[ 525.461716] env[65788]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=65788) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 525.462060] env[65788]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=65788) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 525.462202] env[65788]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=65788) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 525.462546] env[65788]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 525.582950] env[65788]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=65788) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 525.592118] env[65788]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=65788) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 525.637366] env[65788]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 526.196938] env[65788]: INFO nova.virt.driver [None req-f3dfe0ac-edb5-41d0-9395-11bdba673f7d None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 526.267936] env[65788]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 526.268124] env[65788]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 526.268179] env[65788]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=65788) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 529.449162] env[65788]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-3f3cf79c-903a-46b2-b7e9-d024acc7f1e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.466347] env[65788]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=65788) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 529.466527] env[65788]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-7ce2989f-e3c6-40e5-974d-631fc807d43b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.499666] env[65788]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 03c97. [ 529.499842] env[65788]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.232s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 529.500393] env[65788]: INFO nova.virt.vmwareapi.driver [None req-f3dfe0ac-edb5-41d0-9395-11bdba673f7d None None] VMware vCenter version: 7.0.3 [ 529.503826] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fb778b-ae12-44df-9b52-381cc01ee439 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.522097] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0180f5ee-1308-4e9c-8ffb-1065b09e0b31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.528711] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b728a5b6-f9c3-43d7-bcc5-0984cc037ede {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.535826] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6daa8e-8b52-41ec-be0e-04857c24dd2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.549571] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ac4f20-f674-4198-96c4-de2185e87a64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.556203] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6057946-c3aa-497b-81a1-4101fde86ec9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.587694] env[65788]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-09960463-3cec-455f-9c5b-1f9d0f559d83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.594149] env[65788]: DEBUG nova.virt.vmwareapi.driver [None req-f3dfe0ac-edb5-41d0-9395-11bdba673f7d None None] Extension org.openstack.compute already exists. {{(pid=65788) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 529.596898] env[65788]: INFO nova.compute.provider_config [None req-f3dfe0ac-edb5-41d0-9395-11bdba673f7d None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 530.101067] env[65788]: DEBUG nova.context [None req-f3dfe0ac-edb5-41d0-9395-11bdba673f7d None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),e994b8bc-e95d-4c01-a550-60a6b4740f0b(cell1) {{(pid=65788) load_cells /opt/stack/nova/nova/context.py:472}} [ 530.101067] env[65788]: INFO nova.utils [None req-f3dfe0ac-edb5-41d0-9395-11bdba673f7d None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 530.102847] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 530.103245] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 530.104047] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 530.104606] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Acquiring lock "e994b8bc-e95d-4c01-a550-60a6b4740f0b" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 530.104925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Lock "e994b8bc-e95d-4c01-a550-60a6b4740f0b" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 530.106042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Lock "e994b8bc-e95d-4c01-a550-60a6b4740f0b" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 530.128029] env[65788]: INFO dbcounter [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Registered counter for database nova_cell0 [ 530.136187] env[65788]: INFO dbcounter [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Registered counter for database nova_cell1 [ 530.139785] env[65788]: DEBUG oslo_db.sqlalchemy.engines [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65788) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 530.140324] env[65788]: DEBUG oslo_db.sqlalchemy.engines [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=65788) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 530.146017] env[65788]: ERROR nova.db.main.api [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 530.146017] env[65788]: func(*args, **kwargs) [ 530.146017] env[65788]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 530.146017] env[65788]: self.work.run() [ 530.146017] env[65788]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 530.146017] env[65788]: result = self.fn(*self.args, **self.kwargs) [ 530.146017] env[65788]: File "/opt/stack/nova/nova/utils.py", line 584, in context_wrapper [ 530.146017] env[65788]: return func(*args, **kwargs) [ 530.146017] env[65788]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 530.146017] env[65788]: result = fn(*args, **kwargs) [ 530.146017] env[65788]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 530.146017] env[65788]: return f(*args, **kwargs) [ 530.146017] env[65788]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 530.146017] env[65788]: return db.service_get_minimum_version(context, binaries) [ 530.146017] env[65788]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 530.146017] env[65788]: _check_db_access() [ 530.146017] env[65788]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 530.146017] env[65788]: stacktrace = ''.join(traceback.format_stack()) [ 530.146017] env[65788]: [ 530.148392] env[65788]: ERROR nova.db.main.api [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 530.148392] env[65788]: func(*args, **kwargs) [ 530.148392] env[65788]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 530.148392] env[65788]: self.work.run() [ 530.148392] env[65788]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 530.148392] env[65788]: result = self.fn(*self.args, **self.kwargs) [ 530.148392] env[65788]: File "/opt/stack/nova/nova/utils.py", line 584, in context_wrapper [ 530.148392] env[65788]: return func(*args, **kwargs) [ 530.148392] env[65788]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 530.148392] env[65788]: result = fn(*args, **kwargs) [ 530.148392] env[65788]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 530.148392] env[65788]: return f(*args, **kwargs) [ 530.148392] env[65788]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 530.148392] env[65788]: return db.service_get_minimum_version(context, binaries) [ 530.148392] env[65788]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 530.148392] env[65788]: _check_db_access() [ 530.148392] env[65788]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 530.148392] env[65788]: stacktrace = ''.join(traceback.format_stack()) [ 530.148392] env[65788]: [ 530.148392] env[65788]: WARNING nova.objects.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 530.148392] env[65788]: WARNING nova.objects.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Failed to get minimum service version for cell e994b8bc-e95d-4c01-a550-60a6b4740f0b [ 530.148392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Acquiring lock "singleton_lock" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.149376] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Acquired lock "singleton_lock" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 530.149376] env[65788]: DEBUG oslo_concurrency.lockutils [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Releasing lock "singleton_lock" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 530.149376] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Full set of CONF: {{(pid=65788) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:275}} [ 530.149376] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ******************************************************************************** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 530.149376] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] Configuration options gathered from: {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 530.149376] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 530.149639] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 530.149639] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ================================================================================ {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 530.149843] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] allow_resize_to_same_host = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.150141] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] arq_binding_timeout = 300 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.150370] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] backdoor_port = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.150604] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] backdoor_socket = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.150862] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] block_device_allocate_retries = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.151130] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] block_device_allocate_retries_interval = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.151408] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cell_worker_thread_pool_size = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.151686] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cert = self.pem {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.151957] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.152249] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute_monitors = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.152528] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] config_dir = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.152813] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] config_drive_format = iso9660 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.153053] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.153314] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] config_source = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.153590] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] console_host = devstack {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.153889] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] control_exchange = nova {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.154201] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cpu_allocation_ratio = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.154468] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] daemon = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.154758] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] debug = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.155041] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_access_ip_network_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.155307] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_availability_zone = nova {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.155558] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_ephemeral_format = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.155833] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_green_pool_size = 1000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.156181] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.156504] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_schedule_zone = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] default_thread_pool_size = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] disk_allocation_ratio = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] enable_new_services = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] flat_injected = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] force_config_drive = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] force_raw_images = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] graceful_shutdown_timeout = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158298] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] heal_instance_info_cache_interval = -1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158517] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] host = cpu-1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.158812] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] initial_cpu_allocation_ratio = 4.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.159088] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] initial_disk_allocation_ratio = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.159370] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] initial_ram_allocation_ratio = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.159716] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.160057] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_build_timeout = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.160331] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_delete_interval = 300 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.160600] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_format = [instance: %(uuid)s] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.160871] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_name_template = instance-%08x {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.161160] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_usage_audit = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.161433] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_usage_audit_period = month {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.162025] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.162025] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] instances_path = /opt/stack/data/nova/instances {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.162145] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] internal_service_availability_zone = internal {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.162201] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] key = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.162502] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] live_migration_retry_count = 30 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.162776] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_color = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.163062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_config_append = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_dir = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_file = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_options = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_rotate_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_rotate_interval_type = days {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164410] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] log_rotation_type = none {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164410] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164499] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164618] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164774] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.164895] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165058] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] long_rpc_timeout = 1800 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165213] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] max_concurrent_builds = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165362] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] max_concurrent_live_migrations = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165511] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] max_concurrent_snapshots = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165659] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] max_local_block_devices = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165812] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] max_logfile_count = 30 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.165960] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] max_logfile_size_mb = 200 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.166408] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] maximum_instance_delete_attempts = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.166694] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] migrate_max_retries = -1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.166964] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] mkisofs_cmd = genisoimage {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.167309] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] my_block_storage_ip = 10.180.1.21 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.167543] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] my_ip = 10.180.1.21 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168667] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168667] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] network_allocate_retries = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168667] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168667] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] osapi_compute_unique_server_name_scope = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168667] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] password_length = 12 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168667] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] periodic_enable = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168894] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] periodic_fuzzy_delay = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.168967] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] pointer_model = usbtablet {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.169126] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] preallocate_images = none {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.169421] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] publish_errors = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.169680] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] pybasedir = /opt/stack/nova {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.169941] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ram_allocation_ratio = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.170230] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] rate_limit_burst = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.170497] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] rate_limit_except_level = CRITICAL {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.170776] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] rate_limit_interval = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.171051] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reboot_timeout = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.171313] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reclaim_instance_interval = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.171564] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] record = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reimage_timeout_per_gb = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] report_interval = 120 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] rescue_timeout = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reserved_host_cpus = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reserved_host_disk_mb = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reserved_host_memory_mb = 512 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] reserved_huge_pages = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] resize_confirm_window = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] resize_fs_using_block_device = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] resume_guests_state_on_host_boot = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] rpc_response_timeout = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] run_external_periodic_tasks = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] running_deleted_instance_action = reap {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174557] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] running_deleted_instance_poll_interval = 1800 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174557] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] running_deleted_instance_timeout = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.174821] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler_instance_sync_interval = 120 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.175027] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_down_time = 720 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] servicegroup_driver = db {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] shell_completion = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] shelved_offload_time = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] shelved_poll_interval = 3600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] shutdown_timeout = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] source_is_ipv6 = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ssl_only = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] sync_power_state_interval = 600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] sync_power_state_pool_size = 1000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] syslog_log_facility = LOG_USER {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] tempdir = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] thread_pool_statistic_period = -1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] timeout_nbd = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177691] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] transport_url = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177691] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] update_resources_interval = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177747] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] use_cow_images = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.177889] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] use_journal = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.178200] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] use_json = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.178492] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] use_rootwrap_daemon = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.178756] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] use_stderr = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.179061] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] use_syslog = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.179325] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vcpu_pin_set = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.179607] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plugging_is_fatal = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.179880] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plugging_timeout = 300 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.180173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] virt_mkfs = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.180433] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] volume_usage_poll_interval = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.180706] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] watch_log_file = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.180972] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] web = /usr/share/spice-html5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 530.182024] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_concurrency.disable_process_locking = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.182024] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.182182] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.182490] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.182776] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_brick.wait_mpath_device_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.183063] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.183347] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.183623] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_metrics.metrics_process_name = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.183967] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185026] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185026] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.compute_link_prefix = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185026] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185026] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.dhcp_domain = novalocal {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185026] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.enable_instance_password = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185264] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.glance_link_prefix = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185536] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.185823] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.instance_list_cells_batch_strategy = distributed {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.186100] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.instance_list_per_project_cells = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.186380] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.list_records_by_skipping_down_cells = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.186670] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.local_metadata_per_cell = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.186942] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.max_limit = 1000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.metadata_cache_expiration = 15 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.neutron_default_project_id = default {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.response_validation = warn {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.use_neutron_default_nets = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_dynamic_failure_fatal = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.188903] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189190] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_dynamic_ssl_certfile = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189190] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_dynamic_targets = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189190] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_jsonfile_path = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189190] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api.vendordata_providers = ['StaticJSON'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189190] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.backend = dogpile.cache.memcached {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189334] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.backend_argument = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189365] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.backend_expiration_time = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189562] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.config_prefix = cache.oslo {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189731] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.debug_cache_backend = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.189889] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.enable_retry_client = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.190055] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.enable_socket_keepalive = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.190224] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.enabled = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.190379] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.enforce_fips_mode = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.190533] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.expiration_time = 600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.190699] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.hashclient_dead_timeout = 60.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.190852] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.hashclient_retry_attempts = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191019] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.hashclient_retry_timeout = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191174] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_dead_retry = 300 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191320] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_password = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191473] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191626] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191778] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_pool_maxsize = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.191928] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_pool_unused_timeout = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.192090] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_sasl_enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.192262] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_servers = ['localhost:11211'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.192418] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_socket_timeout = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.192566] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.memcache_username = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.192724] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.proxies = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.192885] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_db = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.193057] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_password = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.193225] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_sentinel_service_name = mymaster {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.193392] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.193550] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_server = localhost:6379 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.193708] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_socket_timeout = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.193858] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.redis_username = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.retry_attempts = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194176] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.retry_delay = 0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194328] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.socket_keepalive_count = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194479] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.socket_keepalive_idle = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194632] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.socket_keepalive_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194779] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.tls_allowed_ciphers = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.194924] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.tls_cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.195081] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.tls_certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.195238] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.tls_enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.195383] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cache.tls_keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.195543] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.195707] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.auth_type = password {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.195858] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.196043] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.catalog_info = volumev3::publicURL {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.196201] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.196380] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.196551] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.cross_az_attach = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.196710] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.debug = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.196860] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.endpoint_template = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197026] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.http_retries = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197186] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197336] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197502] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.os_region_name = RegionOne {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197656] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197806] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cinder.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.197967] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.198131] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.cpu_dedicated_set = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.198281] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.cpu_shared_set = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.198458] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.image_type_exclude_list = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.198625] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.live_migration_wait_for_vif_plug = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.198787] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.max_concurrent_disk_ops = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199022] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.max_disk_devices_to_attach = -1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199101] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199262] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199417] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.resource_provider_association_refresh = 300 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199587] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199755] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.shutdown_retry_interval = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.199928] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.200117] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] conductor.workers = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.200285] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] console.allowed_origins = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.200435] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] console.ssl_ciphers = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.200597] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] console.ssl_minimum_version = default {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.200777] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] consoleauth.enforce_session_timeout = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.200948] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] consoleauth.token_ttl = 600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.201138] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.201289] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.201443] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.201596] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.201744] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.201890] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202055] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202209] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202354] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202500] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202649] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202797] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.202943] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.203112] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.service_type = accelerator {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.203268] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.203411] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.203557] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.203705] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.203873] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.204037] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] cyborg.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.204204] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.asyncio_connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.204359] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.asyncio_slave_connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.204546] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.backend = sqlalchemy {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.204735] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.204898] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.connection_debug = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205070] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.connection_parameters = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205230] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.connection_recycle_time = 3600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205385] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.connection_trace = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205539] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.db_inc_retry_interval = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205693] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.db_max_retries = 20 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205844] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.db_max_retry_interval = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.205995] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.db_retry_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.206161] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.max_overflow = 50 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.206315] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.max_pool_size = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.206500] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.max_retries = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.206664] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.mysql_sql_mode = TRADITIONAL {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.206811] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.mysql_wsrep_sync_wait = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.206955] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.pool_timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.207121] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.retry_interval = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.207275] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.slave_connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.207417] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.sqlite_synchronous = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.207587] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] database.use_db_reconnect = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.207775] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.asyncio_connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.207930] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.asyncio_slave_connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.208108] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.backend = sqlalchemy {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.208269] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.208428] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.connection_debug = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.208612] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.connection_parameters = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.208773] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.connection_recycle_time = 3600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.208927] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.connection_trace = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.209095] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.db_inc_retry_interval = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.209251] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.db_max_retries = 20 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.209406] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.db_max_retry_interval = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.209559] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.db_retry_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.209715] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.max_overflow = 50 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.209864] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.max_pool_size = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210023] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.max_retries = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210186] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210334] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.mysql_wsrep_sync_wait = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210482] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.pool_timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210636] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.retry_interval = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210785] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.slave_connection = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.210932] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] api_database.sqlite_synchronous = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.211110] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] devices.enabled_mdev_types = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.211277] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.211435] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ephemeral_storage_encryption.default_format = luks {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.211588] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ephemeral_storage_encryption.enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.211739] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ephemeral_storage_encryption.key_size = 512 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.211898] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.api_servers = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.212059] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.212216] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.212368] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.212514] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.212663] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.212868] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.debug = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213033] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.default_trusted_certificate_ids = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213192] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.enable_certificate_validation = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213343] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.enable_rbd_download = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213489] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213646] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213794] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.213940] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.214095] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.214245] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.num_retries = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.214399] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.rbd_ceph_conf = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.214547] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.rbd_connect_timeout = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.214702] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.rbd_pool = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.214863] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.rbd_user = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215016] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215164] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215310] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215463] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.service_type = image {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215612] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215757] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.215909] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216067] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216236] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216393] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.verify_glance_signatures = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216535] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] glance.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216690] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] guestfs.debug = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216845] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.216996] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.auth_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.217159] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.217306] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.217458] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.217610] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.217760] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.217898] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.218059] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.218241] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.218405] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.218575] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.218725] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.218872] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219023] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219183] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.service_type = shared-file-system {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219336] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.share_apply_policy_timeout = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219490] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219636] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219781] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.219928] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.220110] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.220295] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] manila.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.220469] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] mks.enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.220816] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.220999] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] image_cache.manager_interval = 2400 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.221178] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] image_cache.precache_concurrency = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.221342] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] image_cache.remove_unused_base_images = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.221500] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.221658] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.221820] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] image_cache.subdirectory_name = _base {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.221982] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.api_max_retries = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.222148] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.api_retry_interval = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.222294] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.222443] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.auth_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.222590] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.222762] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.222931] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223098] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.conductor_group = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223249] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223395] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223538] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223688] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223831] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.223974] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.224131] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.224284] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.peer_list = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.224427] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.224575] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.224724] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.serial_console_state_timeout = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.224868] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225031] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.service_type = baremetal {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225179] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.shard = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225329] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225475] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225623] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225770] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.225938] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.226099] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ironic.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.226270] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.226433] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] key_manager.fixed_key = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.226604] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.226753] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.barbican_api_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.226898] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.barbican_endpoint = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227066] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.barbican_endpoint_type = public {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227219] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.barbican_region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227366] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227508] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227658] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227802] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.227947] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.228115] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.number_of_retries = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.228267] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.retry_delay = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.228415] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.send_service_user_token = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.228591] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.228740] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.228889] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.verify_ssl = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229051] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican.verify_ssl_path = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229210] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229362] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.auth_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229506] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229674] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229827] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.229976] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.230134] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.230285] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.230429] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] barbican_service_user.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.230584] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.approle_role_id = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.230727] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.approle_secret_id = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.230881] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.kv_mountpoint = secret {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231043] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.kv_path = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231198] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.kv_version = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231343] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.namespace = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231491] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.root_token_id = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231636] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.ssl_ca_crt_file = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231790] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.timeout = 60.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.231941] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.use_ssl = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.232114] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.232272] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.232420] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.232572] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.232721] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.232867] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233021] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233176] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233321] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233465] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233611] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233754] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.233897] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234053] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234211] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.service_type = identity {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234361] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234507] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234666] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234817] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.234989] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.235147] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] keystone.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.235324] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.ceph_mount_options = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.235635] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.235807] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.connection_uri = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.235962] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_mode = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.236133] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_model_extra_flags = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.236295] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_models = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.236455] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_power_governor_high = performance {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.236613] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_power_governor_low = powersave {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.236768] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_power_management = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.236928] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.237093] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.device_detach_attempts = 8 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.237248] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.device_detach_timeout = 20 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.237402] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.disk_cachemodes = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.237547] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.disk_prefix = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.237700] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.enabled_perf_events = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.237857] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.file_backed_memory = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238020] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.gid_maps = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238174] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.hw_disk_discard = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238321] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.hw_machine_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238510] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_rbd_ceph_conf = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238675] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238829] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.238986] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_rbd_glance_store_name = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.239158] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_rbd_pool = rbd {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.239317] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_type = default {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.239464] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.images_volume_group = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.239642] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.inject_key = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.239805] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.inject_partition = -2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.239957] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.inject_password = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.240125] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.iscsi_iface = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.240280] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.iser_use_multipath = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.240430] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_bandwidth = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.240580] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_completion_timeout = 800 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.240730] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_downtime = 500 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.240918] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_downtime_delay = 75 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241090] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_downtime_steps = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241243] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_inbound_addr = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241395] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_parallel_connections = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241544] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_permit_auto_converge = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241693] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_permit_post_copy = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241838] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_scheme = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.241994] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_timeout_action = abort {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.242159] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_tunnelled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.242305] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_uri = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.242454] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.live_migration_with_native_tls = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.242600] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.max_queues = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.242751] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.mem_stats_period_seconds = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.242974] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.243141] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.nfs_mount_options = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.243437] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.243614] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.num_aoe_discover_tries = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.243773] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.num_iser_scan_tries = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.243925] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.num_memory_encrypted_guests = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.244091] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.num_nvme_discover_tries = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.244246] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.num_pcie_ports = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.244400] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.num_volume_scan_tries = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.244561] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.pmem_namespaces = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.244777] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.quobyte_client_cfg = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.245102] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.245276] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rbd_connect_timeout = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.245437] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.245582] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.245735] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rbd_secret_uuid = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.245884] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rbd_user = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246056] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.realtime_scheduler_priority = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246226] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.remote_filesystem_transport = ssh {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246377] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rescue_image_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246532] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rescue_kernel_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246666] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rescue_ramdisk_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246823] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rng_dev_path = /dev/urandom {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.246968] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.rx_queue_size = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.247138] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.smbfs_mount_options = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.247431] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.247601] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.snapshot_compression = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.247751] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.snapshot_image_format = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.247971] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.248144] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.sparse_logical_volumes = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.248300] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.swtpm_enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.248478] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.swtpm_group = tss {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.248653] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.swtpm_user = tss {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.248813] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.sysinfo_serial = unique {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.248963] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.tb_cache_size = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.249124] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.tx_queue_size = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.249279] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.uid_maps = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.249448] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.use_virtio_for_bridges = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.249617] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.virt_type = kvm {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.249785] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.volume_clear = zero {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.249943] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.volume_clear_size = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.250108] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.volume_enforce_multipath = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.250265] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.volume_use_multipath = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.250411] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_cache_path = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.250567] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.250723] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_mount_group = qemu {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.250878] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_mount_opts = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.251051] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.251337] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.251503] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.vzstorage_mount_user = stack {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.251661] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.251822] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.252028] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.auth_type = password {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.252273] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.252456] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.252619] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.252770] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.252918] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.253089] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.default_floating_pool = public {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.253241] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.253491] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.extension_sync_interval = 600 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.253599] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.http_retries = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.253751] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.253915] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254077] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254286] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.metadata_proxy_shared_secret = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254400] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254542] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.ovs_bridge = br-int {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254694] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.physnets = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254855] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.region_name = RegionOne {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.254996] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.255167] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.service_metadata_proxy = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.255316] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.255468] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.service_type = network {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.255619] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.255763] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.255907] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.256053] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.256225] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.256394] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] neutron.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.256575] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] notifications.bdms_in_notifications = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.256742] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] notifications.default_level = INFO {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.256894] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] notifications.include_share_mapping = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.257067] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] notifications.notification_format = unversioned {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.257225] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] notifications.notify_on_state_change = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.257390] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.257553] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] pci.alias = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.257712] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] pci.device_spec = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.257878] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] pci.report_in_placement = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.258137] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.258266] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.auth_type = password {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.258450] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.auth_url = http://10.180.1.21/identity {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.258604] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.258755] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.258907] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259069] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259221] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259366] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.default_domain_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259515] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.default_domain_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259697] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.domain_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259835] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.domain_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.259980] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.260151] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.260298] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.260443] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.260585] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.260739] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.password = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.260885] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.project_domain_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.261060] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.project_domain_name = Default {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.261220] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.project_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.261385] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.project_name = service {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.261541] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.region_name = RegionOne {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.261699] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.261844] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262015] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.service_type = placement {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262170] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262319] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262465] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262612] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.system_scope = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262759] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.262907] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.trust_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.user_domain_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.user_domain_name = Default {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.user_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.username = nova {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] placement.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264173] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.cores = 20 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264452] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.count_usage_from_placement = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264452] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264452] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.injected_file_content_bytes = 10240 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264599] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.injected_file_path_length = 255 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264755] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.injected_files = 5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.264911] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.instances = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.265078] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.key_pairs = 100 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.265245] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.metadata_items = 128 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.265395] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.ram = 51200 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.265546] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.recheck_quota = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.265702] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.server_group_members = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.265858] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.server_groups = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.266065] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.unified_limits_resource_list = ['servers'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.266242] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] quota.unified_limits_resource_strategy = require {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.266437] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.266608] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.266764] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.image_metadata_prefilter = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.266917] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.267084] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.max_attempts = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.267242] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.max_placement_results = 1000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.267393] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.267543] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.query_placement_for_image_type_support = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.267697] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.267861] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] scheduler.workers = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.268031] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.268199] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.268369] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.268552] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.268723] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.268874] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.269037] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.269226] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.269384] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.host_subset_size = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.269553] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.269717] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.image_properties_default_architecture = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.269883] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.270039] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.image_props_weight_setting = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.270193] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.270347] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.isolated_hosts = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.270496] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.isolated_images = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.270712] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.max_instances_per_host = 50 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.270898] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.271085] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.271248] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.pci_in_placement = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.271402] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.271551] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.271702] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.271854] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.272011] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.272175] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.272327] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.track_instance_changes = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.272496] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.272668] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] metrics.required = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.272852] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] metrics.weight_multiplier = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.273019] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] metrics.weight_of_unavailable = -10000.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.273182] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] metrics.weight_setting = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.273489] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.273654] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] serial_console.enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.273819] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] serial_console.port_range = 10000:20000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.273976] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.274148] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.274308] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] serial_console.serialproxy_port = 6083 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.274463] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.274629] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.auth_type = password {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.274777] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.274925] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.275088] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.275244] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.275389] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.275547] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.send_service_user_token = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.275702] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.275847] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] service_user.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.276027] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.agent_enabled = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.276175] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.276531] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.276709] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.html5proxy_host = 0.0.0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.276872] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.html5proxy_port = 6082 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.277032] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.image_compression = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.277187] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.jpeg_compression = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.277337] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.playback_compression = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.277490] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.require_secure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.277649] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.server_listen = 127.0.0.1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.277802] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.278091] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.278252] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.streaming_mode = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.278405] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] spice.zlib_compression = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.278588] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] upgrade_levels.baseapi = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.278754] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] upgrade_levels.compute = auto {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.278904] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] upgrade_levels.conductor = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.279062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] upgrade_levels.scheduler = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.279223] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.279374] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.auth_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.279524] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.279692] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.279845] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280031] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280148] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280297] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280439] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vendordata_dynamic_auth.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280625] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.api_retry_count = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280778] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.ca_file = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.280937] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.cache_prefix = devstack-image-cache {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.281113] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.cluster_name = testcl1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.281307] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.connection_pool_size = 10 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.281493] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.console_delay_seconds = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.281651] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.datastore_regex = ^datastore.* {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.281893] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.282075] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.host_password = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.282266] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.host_port = 443 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.282440] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.host_username = administrator@vsphere.local {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.282611] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.insecure = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.282751] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.integration_bridge = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.282904] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.maximum_objects = 100 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.pbm_default_policy = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283219] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.pbm_enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283366] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.pbm_wsdl_location = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283522] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283670] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.serial_port_proxy_uri = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283815] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.serial_port_service_uri = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.283969] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.task_poll_interval = 0.5 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.284145] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.use_linked_clone = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.284304] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.vnc_keymap = en-us {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.284458] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.vnc_port = 5900 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.284610] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vmware.vnc_port_total = 10000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.284783] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.auth_schemes = ['none'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.284948] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.285248] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.285427] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.285593] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.novncproxy_port = 6080 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.285773] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.server_listen = 127.0.0.1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.285942] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.286108] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.vencrypt_ca_certs = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.286258] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.vencrypt_client_cert = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.286429] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vnc.vencrypt_client_key = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.286611] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.286770] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.disable_deep_image_inspection = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.286923] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.disable_fallback_pcpu_query = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.287103] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.disable_group_policy_check_upcall = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.287262] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.287413] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.disable_rootwrap = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.287563] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.enable_numa_live_migration = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.287716] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.287865] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288023] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.handle_virt_lifecycle_events = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288183] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.libvirt_disable_apic = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288328] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.never_download_image_if_on_rbd = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288500] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288664] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288814] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.288964] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.289131] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.289283] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.289430] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.289600] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.289856] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.290086] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.290221] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] wsgi.secure_proxy_ssl_header = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.290377] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] zvm.ca_file = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.290527] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] zvm.cloud_connector_url = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.290838] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291014] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] zvm.reachable_timeout = 300 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291183] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291348] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291515] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.connection_string = messaging:// {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291675] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.enabled = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291831] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.es_doc_type = notification {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.291982] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.es_scroll_size = 10000 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.292156] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.es_scroll_time = 2m {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.292308] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.filter_error_trace = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.292461] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.hmac_keys = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.292617] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.sentinel_service_name = mymaster {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.292769] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.socket_timeout = 0.1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.292919] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.trace_requests = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.293081] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler.trace_sqlalchemy = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.293250] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler_jaeger.process_tags = {} {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.293399] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler_jaeger.service_name_prefix = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.293608] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] profiler_otlp.service_name_prefix = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.293788] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.293941] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.294105] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.294258] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.294407] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.294565] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.294718] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.294868] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295033] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295196] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.hostname = devstack {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295343] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295500] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295658] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295812] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.295963] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.296134] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.296284] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.296468] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.296636] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.296787] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.296935] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.297101] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.297253] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.297402] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.297551] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.297700] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.297850] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.298005] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.298165] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.298316] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.298487] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.ssl = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.298715] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.298892] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.299060] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.299223] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.299379] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.ssl_version = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.299540] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.299746] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.299906] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_notifications.retry = -1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.300086] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.300252] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_messaging_notifications.transport_url = **** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.300412] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.auth_section = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.300563] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.auth_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.300712] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.cafile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.300864] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.certfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.301039] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.collect_timing = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.301194] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.connect_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.301341] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.connect_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.301486] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.endpoint_id = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.301665] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.endpoint_interface = publicURL {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.301867] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.endpoint_override = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302035] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.endpoint_region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302188] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.endpoint_service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302336] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.endpoint_service_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302490] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.insecure = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302638] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.keyfile = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302784] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.max_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.302930] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.min_version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303088] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.region_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303238] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.retriable_status_codes = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303384] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.service_name = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303530] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.service_type = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303682] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.split_loggers = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303829] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.status_code_retries = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.303977] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.status_code_retry_delay = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.304138] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.timeout = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.304287] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.valid_interfaces = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.304432] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_limit.version = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.304589] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_reports.file_event_handler = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.304743] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_reports.file_event_handler_interval = 1 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.304890] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] oslo_reports.log_dir = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305064] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305214] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.group = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305358] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305509] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305668] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305815] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.305962] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_linux_bridge_privileged.user = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.306135] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.306283] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.group = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.306454] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.helper_command = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.306615] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.306769] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.306918] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.307083] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] vif_plug_ovs_privileged.user = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.307246] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.flat_interface = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.307411] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.307568] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.307726] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.307882] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.308062] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.308226] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.308376] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_linux_bridge.vlan_interface = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.308568] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.308735] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.isolate_vif = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.308893] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.309066] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.309226] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.309386] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.ovsdb_interface = native {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.309552] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] os_vif_ovs.per_port_bridge = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.309728] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.capabilities = [21, 2] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.309879] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.group = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310036] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.helper_command = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310196] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.log_daemon_traceback = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310350] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310503] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.thread_pool_size = 8 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310651] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] privsep_osbrick.user = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310810] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.310957] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.group = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.311121] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.helper_command = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.311275] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.log_daemon_traceback = False {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.311428] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.311580] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.thread_pool_size = 8 {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.311726] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] nova_sys_admin.user = None {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 530.311846] env[65788]: DEBUG oslo_service.backend._eventlet.service [None req-684fd928-9cc6-4a16-bfc1-bb0598d66fae None None] ******************************************************************************** {{(pid=65788) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 530.312279] env[65788]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 530.816277] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Getting list of instances from cluster (obj){ [ 530.816277] env[65788]: value = "domain-c8" [ 530.816277] env[65788]: _type = "ClusterComputeResource" [ 530.816277] env[65788]: } {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 530.817618] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c4fbfa-9a13-4645-aee4-eb4b52c7ddb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.827730] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Got total of 0 instances {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 530.828332] env[65788]: WARNING nova.virt.vmwareapi.driver [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 530.829020] env[65788]: INFO nova.virt.node [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Generated node identity 3a6417f0-b161-4086-8a0e-1cb139eae377 [ 530.829266] env[65788]: INFO nova.virt.node [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Wrote node identity 3a6417f0-b161-4086-8a0e-1cb139eae377 to /opt/stack/data/n-cpu-1/compute_id [ 531.331930] env[65788]: WARNING nova.compute.manager [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Compute nodes ['3a6417f0-b161-4086-8a0e-1cb139eae377'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 532.337967] env[65788]: INFO nova.compute.manager [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 533.344641] env[65788]: WARNING nova.compute.manager [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 533.345033] env[65788]: DEBUG oslo_concurrency.lockutils [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 533.345179] env[65788]: DEBUG oslo_concurrency.lockutils [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 533.345320] env[65788]: DEBUG oslo_concurrency.lockutils [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 533.345533] env[65788]: DEBUG nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 533.346538] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16160747-1dd1-4b8b-bb7d-74e84228843a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.355543] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75139ca-c83a-4ba2-ad11-6a4f11fc55c6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.371913] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c0adb6-cd93-461a-a1d7-0d5b98026382 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.379130] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518b5f9d-910a-4ad1-b5cd-386b10e2a460 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.408854] env[65788]: DEBUG nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180681MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 533.409025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 533.409215] env[65788]: DEBUG oslo_concurrency.lockutils [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 533.911849] env[65788]: WARNING nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] No compute node record for cpu-1:3a6417f0-b161-4086-8a0e-1cb139eae377: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 3a6417f0-b161-4086-8a0e-1cb139eae377 could not be found. [ 534.416130] env[65788]: INFO nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 3a6417f0-b161-4086-8a0e-1cb139eae377 [ 535.924507] env[65788]: DEBUG nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 535.924921] env[65788]: DEBUG nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 536.095416] env[65788]: INFO nova.scheduler.client.report [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] [req-1b38c5ef-9e02-46ac-83ca-c6f5a9e0b155] Created resource provider record via placement API for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 536.113508] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4047d263-5263-46b5-b606-9cfeddbd0fd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.121929] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7e3b3c-dea5-4c1e-9a20-c6fa6293fc90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.153556] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2875aea1-6e08-4df2-bf52-556760616249 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.162063] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d648338-cd60-4d04-869c-a0f0f1ba72b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.176442] env[65788]: DEBUG nova.compute.provider_tree [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 536.716053] env[65788]: DEBUG nova.scheduler.client.report [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 536.716342] env[65788]: DEBUG nova.compute.provider_tree [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 0 to 1 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 536.716500] env[65788]: DEBUG nova.compute.provider_tree [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 536.780785] env[65788]: DEBUG nova.compute.provider_tree [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 1 to 2 during operation: update_traits {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 537.285426] env[65788]: DEBUG nova.compute.resource_tracker [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 537.285845] env[65788]: DEBUG oslo_concurrency.lockutils [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.876s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 537.285845] env[65788]: DEBUG nova.service [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Creating RPC server for service compute {{(pid=65788) start /opt/stack/nova/nova/service.py:177}} [ 537.300593] env[65788]: DEBUG nova.service [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] Join ServiceGroup membership for this service compute {{(pid=65788) start /opt/stack/nova/nova/service.py:194}} [ 537.300840] env[65788]: DEBUG nova.servicegroup.drivers.db [None req-11a4fefa-3379-44a5-a988-4621e9ddb4bc None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=65788) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 541.304101] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_power_states {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 541.806528] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Getting list of instances from cluster (obj){ [ 541.806528] env[65788]: value = "domain-c8" [ 541.806528] env[65788]: _type = "ClusterComputeResource" [ 541.806528] env[65788]: } {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 541.807706] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507399f4-ae79-4a99-82e7-4c5427a55caf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.816866] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Got total of 0 instances {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 541.817105] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 541.817415] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Getting list of instances from cluster (obj){ [ 541.817415] env[65788]: value = "domain-c8" [ 541.817415] env[65788]: _type = "ClusterComputeResource" [ 541.817415] env[65788]: } {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 541.818299] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da3ee2c-af87-4243-b462-6b72aea10d4c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.825861] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Got total of 0 instances {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 580.835168] env[65788]: INFO nova.utils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The default thread pool MainProcess.default is initialized [ 580.837544] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "48ac772b-3b3b-403e-814d-b5290b17fcad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 580.837544] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 581.342801] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 581.904879] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 581.905413] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 581.907510] env[65788]: INFO nova.compute.claims [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.958843] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4beaa8a-d6e0-4785-a46f-de3bba8e1bd7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.968582] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a2458d-bd6e-45fa-be26-9693ee809460 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.012815] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5b5896-edc4-4165-9c0b-6a024cc7cfeb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.022608] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659c449d-dfbe-4292-9da6-e445d7b61bd3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.039731] env[65788]: DEBUG nova.compute.provider_tree [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.544152] env[65788]: DEBUG nova.scheduler.client.report [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 584.021295] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "ed55713e-8bf1-4960-8309-47498b163d02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 584.021673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 584.058031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.150s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 584.058031] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 584.394874] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "39127362-8daf-43ff-99cb-ddb54c6fe095" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 584.394874] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "39127362-8daf-43ff-99cb-ddb54c6fe095" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 584.528951] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 584.562161] env[65788]: DEBUG nova.compute.utils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 584.564404] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 584.564792] env[65788]: DEBUG nova.network.neutron [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 584.571020] env[65788]: WARNING neutronclient.v2_0.client [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 584.571020] env[65788]: WARNING neutronclient.v2_0.client [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 584.572305] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 584.573224] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 584.900249] env[65788]: DEBUG nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 585.058893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 585.058893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 585.060155] env[65788]: INFO nova.compute.claims [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.083741] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 585.239520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 585.240848] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 585.425482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 585.646738] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.647142] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.647369] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.647562] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.647746] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.647927] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.648151] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.648293] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 585.648731] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 585.742982] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 586.094482] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 586.125894] env[65788]: DEBUG nova.policy [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a7ca4f0daa04471b948f6affd617df3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1652d71008d45b28f4139d6edd8bf2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 586.139260] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 586.139534] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 586.139696] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 586.139872] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 586.140099] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 586.140236] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 586.140525] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.140695] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 586.141322] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 586.141513] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 586.141688] env[65788]: DEBUG nova.virt.hardware [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 586.144669] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b8d037-62c6-43ef-9e77-3ead4ff9a684 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.154870] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 586.156752] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf805c38-770d-435e-9700-70af5172d499 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.177217] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2682574c-7110-4561-ae19-e9dcc607145d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.246977] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898120fb-d972-418a-8058-3643a9b2a543 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.259047] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfda8a5-21d1-4b29-8198-dc6b760c2eb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.297233] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 586.297233] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700e5cde-300d-4293-8ef6-54711b79cd50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.306189] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3719f068-72d8-48db-bb6d-9595e0d807fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.322143] env[65788]: DEBUG nova.compute.provider_tree [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.827033] env[65788]: DEBUG nova.scheduler.client.report [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 587.334725] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 587.336120] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 587.340244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.914s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 587.341751] env[65788]: INFO nova.compute.claims [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.560624] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 587.560925] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 587.691754] env[65788]: DEBUG nova.network.neutron [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Successfully created port: 4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 587.848249] env[65788]: DEBUG nova.compute.utils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 587.848990] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 587.853102] env[65788]: DEBUG nova.network.neutron [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 587.853522] env[65788]: WARNING neutronclient.v2_0.client [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 587.854431] env[65788]: WARNING neutronclient.v2_0.client [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 587.854431] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 587.854899] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 588.025933] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "c06991cd-15b0-4d48-849a-140720ee07a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.026168] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "c06991cd-15b0-4d48-849a-140720ee07a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 588.351888] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "4bedb939-e86b-42bd-a490-a73086ecfd7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 588.351888] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 588.368215] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 588.490020] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2472e503-61c1-458f-8d7c-9d2111569c13 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.501532] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7774a8ac-6e27-4743-b6d2-968217897a8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.535759] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 588.541184] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd400ec9-f49c-4668-ba38-c548e117b324 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.549236] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a4730a-1c62-4a52-9212-83e08b5f015a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.566727] env[65788]: DEBUG nova.compute.provider_tree [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.614944] env[65788]: DEBUG nova.policy [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bedb4502cfc74d76b0c6df2d9173855e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a38ce4af9b414c778a69cd3f64f5f6bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 588.854461] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 589.071480] env[65788]: DEBUG nova.scheduler.client.report [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 589.084073] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.208914] env[65788]: DEBUG nova.network.neutron [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Successfully created port: ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 589.379540] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 589.386398] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 589.410840] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 589.411119] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 589.411270] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 589.411442] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 589.412492] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 589.413185] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 589.413185] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.413185] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 589.413324] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 589.413570] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 589.413631] env[65788]: DEBUG nova.virt.hardware [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 589.417065] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cac213a-67b0-42cf-ab76-c3cdcbf28802 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.426338] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cca44a0-a2c8-4b61-b54a-9ac732db2d8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.514031] env[65788]: DEBUG nova.network.neutron [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Successfully updated port: 4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 589.581495] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.240s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 589.581495] env[65788]: DEBUG nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 589.583346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.429s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 589.583592] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 589.583700] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 589.583938] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.288s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 589.585342] env[65788]: INFO nova.compute.claims [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.589134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40abd438-37f4-41fb-9c32-1172988cc87e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.598288] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2767eb3e-61c4-44ec-b53d-f23145660082 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.615134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aef4e14-b3ab-4341-a993-43439bfbf549 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.625461] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8373508-3598-4fc4-8523-d4519da6e551 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.666470] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180690MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 589.666627] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 590.016996] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "refresh_cache-48ac772b-3b3b-403e-814d-b5290b17fcad" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.017213] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired lock "refresh_cache-48ac772b-3b3b-403e-814d-b5290b17fcad" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 590.017934] env[65788]: DEBUG nova.network.neutron [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 590.090344] env[65788]: DEBUG nova.compute.utils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 590.091781] env[65788]: DEBUG nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 590.523383] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 590.523383] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 590.595503] env[65788]: DEBUG nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 590.738199] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfaba21-798f-49df-aa09-f92d90ee9c5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.747077] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5c4674-11bf-431e-bf69-6d8158a33608 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.777652] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffba1f8-a25c-4a14-830a-aea1780d58d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.785648] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4680441-3952-482a-9683-14612d1a8d0f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.800101] env[65788]: DEBUG nova.compute.provider_tree [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.112927] env[65788]: DEBUG nova.network.neutron [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Successfully updated port: ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 591.185518] env[65788]: DEBUG nova.network.neutron [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 591.303068] env[65788]: DEBUG nova.scheduler.client.report [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 591.473172] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 591.473592] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 591.613056] env[65788]: DEBUG nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 591.616243] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.616585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 591.617083] env[65788]: DEBUG nova.network.neutron [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 591.661478] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 591.661478] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 591.661478] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 591.662104] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 591.662104] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 591.662104] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 591.662104] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.662104] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 591.662250] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 591.662250] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 591.663092] env[65788]: DEBUG nova.virt.hardware [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 591.663983] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a893ed-c2a6-4967-8f60-7dfe3182874d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.678036] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3909d7ab-58e7-4a96-913d-f1e49550f678 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.700015] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 591.713802] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 591.715041] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e2f4f80-d950-47dc-a0f2-31238727a88c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.730884] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Created folder: OpenStack in parent group-v4. [ 591.731124] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Creating folder: Project (879e3dd0f9d3415a88ac799eb9669f08). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 591.732192] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-309be37f-c8f6-4e71-b05b-76cb0fc06b8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.742420] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Created folder: Project (879e3dd0f9d3415a88ac799eb9669f08) in parent group-v910111. [ 591.744077] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Creating folder: Instances. Parent ref: group-v910112. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 591.744077] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ff1c76a-0696-4186-bffc-5f6680d0be25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.755916] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Created folder: Instances in parent group-v910112. [ 591.756259] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 591.756477] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 591.756699] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66ceec92-a97f-4f11-9a85-8f7aa52157a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.776882] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 591.776882] env[65788]: value = "task-4661666" [ 591.776882] env[65788]: _type = "Task" [ 591.776882] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.789391] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661666, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.809586] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.225s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 591.810143] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 591.815688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.732s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 591.817329] env[65788]: INFO nova.compute.claims [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.119794] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 592.120216] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 592.256531] env[65788]: WARNING neutronclient.v2_0.client [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.256531] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 592.256531] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 592.293298] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661666, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.326904] env[65788]: DEBUG nova.compute.utils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 592.329678] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 592.330177] env[65788]: DEBUG nova.network.neutron [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 592.330707] env[65788]: WARNING neutronclient.v2_0.client [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.331326] env[65788]: WARNING neutronclient.v2_0.client [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 592.331687] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 592.332081] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 592.341789] env[65788]: DEBUG nova.network.neutron [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 592.790548] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661666, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.833620] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 592.834570] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 592.848034] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 592.908197] env[65788]: DEBUG nova.network.neutron [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Updating instance_info_cache with network_info: [{"id": "4dd6a49e-86fd-442c-bd8d-4acf8587707f", "address": "fa:16:3e:8a:b7:d2", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dd6a49e-86", "ovs_interfaceid": "4dd6a49e-86fd-442c-bd8d-4acf8587707f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 592.968609] env[65788]: DEBUG nova.compute.manager [req-de5950c6-d9f3-43a5-97aa-ebc2cf477e90 req-464a6028-ba21-4e72-b2c3-9009554b91aa service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Received event network-vif-plugged-4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 592.968609] env[65788]: DEBUG oslo_concurrency.lockutils [req-de5950c6-d9f3-43a5-97aa-ebc2cf477e90 req-464a6028-ba21-4e72-b2c3-9009554b91aa service nova] Acquiring lock "48ac772b-3b3b-403e-814d-b5290b17fcad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 592.969429] env[65788]: DEBUG oslo_concurrency.lockutils [req-de5950c6-d9f3-43a5-97aa-ebc2cf477e90 req-464a6028-ba21-4e72-b2c3-9009554b91aa service nova] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 592.969429] env[65788]: DEBUG oslo_concurrency.lockutils [req-de5950c6-d9f3-43a5-97aa-ebc2cf477e90 req-464a6028-ba21-4e72-b2c3-9009554b91aa service nova] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 592.969429] env[65788]: DEBUG nova.compute.manager [req-de5950c6-d9f3-43a5-97aa-ebc2cf477e90 req-464a6028-ba21-4e72-b2c3-9009554b91aa service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] No waiting events found dispatching network-vif-plugged-4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 592.969616] env[65788]: WARNING nova.compute.manager [req-de5950c6-d9f3-43a5-97aa-ebc2cf477e90 req-464a6028-ba21-4e72-b2c3-9009554b91aa service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Received unexpected event network-vif-plugged-4dd6a49e-86fd-442c-bd8d-4acf8587707f for instance with vm_state building and task_state spawning. [ 593.028205] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "141aefe8-1b95-4963-854d-da79ddf143f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 593.028437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "141aefe8-1b95-4963-854d-da79ddf143f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 593.061629] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fd1b2e-ec4b-4bbf-a131-003f821533e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.071997] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0841df6e-7c0d-4a0d-8782-7dd9b52e21fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.108104] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f940f5c4-6b34-4919-9760-1c016f927509 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.117397] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6173665d-522d-428a-8177-7184f3fb2118 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.136463] env[65788]: DEBUG nova.compute.provider_tree [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.291350] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661666, 'name': CreateVM_Task, 'duration_secs': 1.334338} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.291578] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 593.293240] env[65788]: DEBUG oslo_vmware.service [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02bb768-32ab-4af7-a578-e0b70bfa9fd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.302559] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.303844] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 593.303844] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 593.303844] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66fb4ea3-37a7-4e53-89b8-87518a003371 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.310540] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 593.310540] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fcee53-d025-8192-c02b-c1d3ed4bbfd0" [ 593.310540] env[65788]: _type = "Task" [ 593.310540] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.318149] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fcee53-d025-8192-c02b-c1d3ed4bbfd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.411514] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Releasing lock "refresh_cache-48ac772b-3b3b-403e-814d-b5290b17fcad" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 593.412353] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Instance network_info: |[{"id": "4dd6a49e-86fd-442c-bd8d-4acf8587707f", "address": "fa:16:3e:8a:b7:d2", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dd6a49e-86", "ovs_interfaceid": "4dd6a49e-86fd-442c-bd8d-4acf8587707f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 593.413390] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:b7:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4dd6a49e-86fd-442c-bd8d-4acf8587707f', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 593.422242] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Creating folder: Project (b1652d71008d45b28f4139d6edd8bf2f). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 593.423853] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-661cfe86-ef86-420e-a50a-c751d332e95d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.435350] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Created folder: Project (b1652d71008d45b28f4139d6edd8bf2f) in parent group-v910111. [ 593.435557] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Creating folder: Instances. Parent ref: group-v910115. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 593.435812] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7d8513a-3446-4bac-9b0b-b9e59cb0c5af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.449241] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Created folder: Instances in parent group-v910115. [ 593.450157] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 593.450157] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 593.450157] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ba2eb29-2c2d-4725-9bf0-909fc57f7814 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.474272] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 593.474272] env[65788]: value = "task-4661669" [ 593.474272] env[65788]: _type = "Task" [ 593.474272] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.483217] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661669, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.487871] env[65788]: DEBUG nova.policy [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de952e5205da495096409b2f66195739', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d3dfcc412774144893f0e10a828848c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 593.532927] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 593.639841] env[65788]: DEBUG nova.scheduler.client.report [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 593.821517] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 593.821812] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 593.822156] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.822344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 593.822859] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 593.823195] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fea64422-6071-4e3d-82da-884e5911c9ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.843213] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 593.843373] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 593.844838] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f51ee2a-ec61-49d8-8735-b62cc16d16b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.852394] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad474048-b043-438b-a83e-5b2dab08cf86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.858470] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 593.858470] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d1893c-5089-2faa-2874-d705b14e06bd" [ 593.858470] env[65788]: _type = "Task" [ 593.858470] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.868127] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 593.870178] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d1893c-5089-2faa-2874-d705b14e06bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.890018] env[65788]: WARNING neutronclient.v2_0.client [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 593.890809] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 593.891164] env[65788]: WARNING openstack [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 593.909947] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 593.910335] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 593.910418] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 593.910533] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 593.910674] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 593.910905] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 593.911030] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.911252] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 593.911352] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 593.911509] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 593.911677] env[65788]: DEBUG nova.virt.hardware [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 593.912890] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de97d02e-e351-46d5-85e6-5324be848dc5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.922357] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672ebeb6-7307-46d9-83d8-af8415185ccc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.984804] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661669, 'name': CreateVM_Task, 'duration_secs': 0.380234} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.984934] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 593.985452] env[65788]: WARNING neutronclient.v2_0.client [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 593.986354] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 593.987303] env[65788]: WARNING openstack [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.067778] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 594.146838] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 594.147839] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 594.152605] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.766s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 594.155043] env[65788]: INFO nova.compute.claims [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 594.203589] env[65788]: DEBUG nova.network.neutron [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Successfully created port: 5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 594.325941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.326154] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 594.328964] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 594.328964] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e71e4e8e-8d2c-456e-ac0d-5a878cffba35 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.336874] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 594.336874] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c196cb-e5ec-bd7a-9dd5-dc848f4183ab" [ 594.336874] env[65788]: _type = "Task" [ 594.336874] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.344245] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c196cb-e5ec-bd7a-9dd5-dc848f4183ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.370426] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 594.370772] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Creating directory with path [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 594.371090] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51ceba1a-1bcb-43d9-8a7f-50fe00a5118b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.398132] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Created directory with path [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 594.398328] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Fetch image to [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 594.398503] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Downloading image file data d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk on the data store datastore2 {{(pid=65788) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 594.399363] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f83415-47c8-4dd3-87ba-3f8481b5ad21 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.409344] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab8195d-551d-4b5e-9cc3-8a155111d84e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.420138] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c5b294-c865-4b0a-9c1d-196f668254ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.452751] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e701cd2-90d9-4a04-a272-5c9b2f58a915 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.460525] env[65788]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-ae521309-2206-48a5-abd2-166b87d4073c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.496363] env[65788]: DEBUG nova.network.neutron [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [{"id": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "address": "fa:16:3e:2c:93:04", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa53d50-ad", "ovs_interfaceid": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 594.504054] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Downloading image file data d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to the data store datastore2 {{(pid=65788) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 594.585934] env[65788]: DEBUG oslo_vmware.rw_handles [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 594.661117] env[65788]: DEBUG nova.compute.utils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 594.664311] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 594.665622] env[65788]: DEBUG nova.network.neutron [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 594.665622] env[65788]: WARNING neutronclient.v2_0.client [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 594.665622] env[65788]: WARNING neutronclient.v2_0.client [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 594.665914] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 594.666105] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 594.847451] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 594.848072] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 594.848072] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.980461] env[65788]: DEBUG nova.policy [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '500bd0aefd47405fa6e17a866543a19b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4bce3aed585840338789b5058627b1a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 595.002825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 595.003317] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Instance network_info: |[{"id": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "address": "fa:16:3e:2c:93:04", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa53d50-ad", "ovs_interfaceid": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 595.005165] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:93:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ffa53d50-ad0e-438c-994a-b0b7a97f85f2', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 595.013602] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Creating folder: Project (a38ce4af9b414c778a69cd3f64f5f6bf). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 595.020022] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d0f3d87-be98-4c0f-9467-6eee826b4b46 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.032479] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Created folder: Project (a38ce4af9b414c778a69cd3f64f5f6bf) in parent group-v910111. [ 595.032923] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Creating folder: Instances. Parent ref: group-v910118. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 595.034171] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-825e54ac-8192-4476-aeb7-b6321837c444 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.047077] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Created folder: Instances in parent group-v910118. [ 595.047077] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 595.047302] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 595.047616] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a72415f-c0bf-4ba0-b659-848de8076853 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.073819] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 595.073819] env[65788]: value = "task-4661672" [ 595.073819] env[65788]: _type = "Task" [ 595.073819] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.084757] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661672, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.180781] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 595.232631] env[65788]: DEBUG oslo_vmware.rw_handles [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 595.232870] env[65788]: DEBUG oslo_vmware.rw_handles [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 595.330947] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd072ac-2e2b-475f-a0ca-b468cd8316a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.340027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb90a80-cf7d-4aec-bd59-ee50f4a8ef93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.375235] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf54d2-197c-414f-af82-8585acd28917 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.381288] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Downloaded image file data d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk on the data store datastore2 {{(pid=65788) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 595.383141] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 595.383826] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Copying Virtual Disk [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk to [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 595.383826] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd2e2cc3-445a-45cf-8063-411617fab362 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.391984] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b40405-11e7-4ab0-b613-2af8315899f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.411117] env[65788]: DEBUG nova.compute.provider_tree [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.416304] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 595.416304] env[65788]: value = "task-4661673" [ 595.416304] env[65788]: _type = "Task" [ 595.416304] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.424701] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661673, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.470610] env[65788]: DEBUG nova.network.neutron [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Successfully created port: 752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 595.586317] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661672, 'name': CreateVM_Task, 'duration_secs': 0.40683} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.586500] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 595.587065] env[65788]: WARNING neutronclient.v2_0.client [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 595.587835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.587835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 595.587835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 595.588102] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6239dcf-7065-453b-b78f-ff00271938bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.593510] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 595.593510] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d70180-e742-5415-50df-45157dbec0b4" [ 595.593510] env[65788]: _type = "Task" [ 595.593510] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.601900] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d70180-e742-5415-50df-45157dbec0b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.916221] env[65788]: DEBUG nova.scheduler.client.report [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 595.934839] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661673, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.044316] env[65788]: DEBUG nova.network.neutron [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Successfully updated port: 5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 596.109284] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 596.109392] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 596.109722] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.194912] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 596.220664] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 596.221045] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 596.221117] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 596.221252] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 596.221394] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 596.221538] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 596.221749] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.221899] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 596.222085] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 596.222255] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 596.222424] env[65788]: DEBUG nova.virt.hardware [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 596.223321] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995081d4-c946-48fe-89db-adf6d6bc6771 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.233830] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516bc146-152e-488a-90ee-2d78909aa1b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.423909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 596.424493] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 596.427106] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.760s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 596.439747] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661673, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693202} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.440269] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Copied Virtual Disk [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk to [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 596.440269] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleting the datastore file [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 596.440443] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62893060-2d6a-411c-9ebf-9532f056ddb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.449281] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 596.449281] env[65788]: value = "task-4661674" [ 596.449281] env[65788]: _type = "Task" [ 596.449281] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.457248] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.546686] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "refresh_cache-98a2efec-cb20-4e95-9ed5-5b2057a48c39" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.546881] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquired lock "refresh_cache-98a2efec-cb20-4e95-9ed5-5b2057a48c39" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 596.547072] env[65788]: DEBUG nova.network.neutron [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 596.940564] env[65788]: DEBUG nova.compute.utils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 596.942993] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 596.942993] env[65788]: DEBUG nova.network.neutron [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 596.942993] env[65788]: WARNING neutronclient.v2_0.client [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.943232] env[65788]: WARNING neutronclient.v2_0.client [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 596.943932] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 596.944124] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 596.964237] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02674} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.964524] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 596.964798] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Moving file from [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b. {{(pid=65788) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 596.965051] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-557f5596-6905-4bd5-a668-6f9c272843a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.973324] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 596.973324] env[65788]: value = "task-4661675" [ 596.973324] env[65788]: _type = "Task" [ 596.973324] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.983644] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661675, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.992160] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "efe1048b-50e9-4add-910a-607a95759c7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 596.992398] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "efe1048b-50e9-4add-910a-607a95759c7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 597.054063] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 597.054063] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 597.453603] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 597.474413] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 48ac772b-3b3b-403e-814d-b5290b17fcad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 597.474525] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance ed55713e-8bf1-4960-8309-47498b163d02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 597.474660] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 39127362-8daf-43ff-99cb-ddb54c6fe095 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 597.474821] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 98a2efec-cb20-4e95-9ed5-5b2057a48c39 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 597.474934] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance c06991cd-15b0-4d48-849a-140720ee07a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 597.475075] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4bedb939-e86b-42bd-a490-a73086ecfd7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 597.496535] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 597.499362] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661675, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030017} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.499837] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] File moved {{(pid=65788) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 597.502031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Cleaning up location [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4 {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 597.502031] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleting the datastore file [datastore2] vmware_temp/8a5ed7e6-3d75-4455-947d-93f4aaeef7b4 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 597.502031] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-216d5afb-dcc8-4724-9425-3915818eaa37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.511919] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 597.511919] env[65788]: value = "task-4661676" [ 597.511919] env[65788]: _type = "Task" [ 597.511919] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.522479] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.894676] env[65788]: DEBUG nova.policy [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eda3217367c04a2dab30bb8ae2a97a2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'becee551c9054443a78c0bc2765b3a17', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 597.925955] env[65788]: DEBUG nova.network.neutron [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Successfully updated port: 752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 597.932868] env[65788]: DEBUG nova.network.neutron [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 597.986310] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 141aefe8-1b95-4963-854d-da79ddf143f7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 598.013534] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.014037] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.032644] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.036939] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028483} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.037258] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 598.038241] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d179e3c3-0a34-4adb-ac02-d09dcc5e03b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.045975] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 598.045975] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52da90a5-a79e-3a10-077c-dbf08d4618f4" [ 598.045975] env[65788]: _type = "Task" [ 598.045975] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.056749] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52da90a5-a79e-3a10-077c-dbf08d4618f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.179653] env[65788]: WARNING neutronclient.v2_0.client [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 598.179653] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.179653] env[65788]: WARNING openstack [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.332410] env[65788]: DEBUG nova.network.neutron [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Successfully created port: a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 598.374933] env[65788]: DEBUG nova.network.neutron [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Updating instance_info_cache with network_info: [{"id": "5cc1dfe1-7bda-4f6a-9828-254a5f93a508", "address": "fa:16:3e:91:98:45", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cc1dfe1-7b", "ovs_interfaceid": "5cc1dfe1-7bda-4f6a-9828-254a5f93a508", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 598.433134] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "refresh_cache-c06991cd-15b0-4d48-849a-140720ee07a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.433134] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquired lock "refresh_cache-c06991cd-15b0-4d48-849a-140720ee07a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 598.433648] env[65788]: DEBUG nova.network.neutron [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 598.469181] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 598.494616] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 598.494616] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 598.494616] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=100GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '6', 'num_vm_building': '6', 'num_task_spawning': '5', 'num_os_type_None': '6', 'num_proj_b1652d71008d45b28f4139d6edd8bf2f': '1', 'io_workload': '6', 'num_proj_a38ce4af9b414c778a69cd3f64f5f6bf': '1', 'num_proj_879e3dd0f9d3415a88ac799eb9669f08': '1', 'num_proj_6d3dfcc412774144893f0e10a828848c': '1', 'num_proj_4bce3aed585840338789b5058627b1a8': '1', 'num_task_None': '1', 'num_proj_becee551c9054443a78c0bc2765b3a17': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 598.498892] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 598.499163] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 598.499319] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 598.499530] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 598.499676] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 598.499818] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 598.500027] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.500228] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 598.500439] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 598.500782] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 598.500973] env[65788]: DEBUG nova.virt.hardware [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 598.502384] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0d2992-18da-4e01-a67d-f5908e56d4b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.512726] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f295566-9ee2-4fde-a16d-499d34008189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.560308] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52da90a5-a79e-3a10-077c-dbf08d4618f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009789} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.560465] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 598.560747] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 39127362-8daf-43ff-99cb-ddb54c6fe095/39127362-8daf-43ff-99cb-ddb54c6fe095.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 598.561189] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 598.561230] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 598.561419] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ed93e8b-c5b3-4b68-b6aa-f10103667b12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.564291] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-769ffc49-940c-4f1a-bce6-7e9eb4913eee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.576984] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 598.576984] env[65788]: value = "task-4661677" [ 598.576984] env[65788]: _type = "Task" [ 598.576984] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.580274] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 598.580274] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 598.582803] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09da112a-9acf-4d7d-8f4e-686b3e30f630 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.592883] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.594807] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 598.594807] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c54ee3-c4d5-ed34-6ff8-1a630b533bb9" [ 598.594807] env[65788]: _type = "Task" [ 598.594807] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.609429] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c54ee3-c4d5-ed34-6ff8-1a630b533bb9, 'name': SearchDatastore_Task, 'duration_secs': 0.0104} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.612939] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-195d8064-a4e4-44a3-8760-335b3d407203 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.620400] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 598.620400] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52668bc8-5f9d-3419-c7d7-295a367bad21" [ 598.620400] env[65788]: _type = "Task" [ 598.620400] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.629124] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52668bc8-5f9d-3419-c7d7-295a367bad21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.675772] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a487b37-edd7-4fac-8d2d-e6f5f7e57965 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.685054] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc43701-77ce-41fd-8987-4467a583ee47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.725024] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b29cf88-63b5-48c0-b07e-5d0c120fa49e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.733272] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7ca5b0-cdb9-4ac8-b51e-7b766c5dd823 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.749886] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.863396] env[65788]: DEBUG nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Received event network-changed-4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 598.863396] env[65788]: DEBUG nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Refreshing instance network info cache due to event network-changed-4dd6a49e-86fd-442c-bd8d-4acf8587707f. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 598.863396] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Acquiring lock "refresh_cache-48ac772b-3b3b-403e-814d-b5290b17fcad" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.863535] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Acquired lock "refresh_cache-48ac772b-3b3b-403e-814d-b5290b17fcad" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 598.863664] env[65788]: DEBUG nova.network.neutron [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Refreshing network info cache for port 4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 598.878403] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Releasing lock "refresh_cache-98a2efec-cb20-4e95-9ed5-5b2057a48c39" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 598.878804] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Instance network_info: |[{"id": "5cc1dfe1-7bda-4f6a-9828-254a5f93a508", "address": "fa:16:3e:91:98:45", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cc1dfe1-7b", "ovs_interfaceid": "5cc1dfe1-7bda-4f6a-9828-254a5f93a508", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 598.883061] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:98:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5cc1dfe1-7bda-4f6a-9828-254a5f93a508', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 598.894662] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Creating folder: Project (6d3dfcc412774144893f0e10a828848c). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 598.894662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 598.894662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 598.894662] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aef0940b-4f73-4a93-add2-0166dec5b1a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.908679] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Created folder: Project (6d3dfcc412774144893f0e10a828848c) in parent group-v910111. [ 598.908679] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Creating folder: Instances. Parent ref: group-v910121. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 598.909497] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca9c0af6-c6b6-46ae-8044-3bf6f145a52b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.921743] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Created folder: Instances in parent group-v910121. [ 598.922104] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 598.922353] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 598.922564] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2af9815-cf12-4cec-98df-480e8a5c32c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.939395] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 598.939879] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 598.959332] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 598.959332] env[65788]: value = "task-4661680" [ 598.959332] env[65788]: _type = "Task" [ 598.959332] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.972035] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661680, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.028214] env[65788]: DEBUG nova.network.neutron [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 599.078797] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.079609] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.101293] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661677, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.133427] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52668bc8-5f9d-3419-c7d7-295a367bad21, 'name': SearchDatastore_Task, 'duration_secs': 0.010969} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.133427] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 599.133427] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 48ac772b-3b3b-403e-814d-b5290b17fcad/48ac772b-3b3b-403e-814d-b5290b17fcad.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 599.133427] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 599.134258] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 599.134258] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd7871a5-8e17-4f8c-bc62-4de8493f6819 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.134825] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5d814ae-38f8-46f3-9953-4e7ce5514c1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.143540] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 599.143540] env[65788]: value = "task-4661681" [ 599.143540] env[65788]: _type = "Task" [ 599.143540] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.149482] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 599.149692] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 599.153759] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-622bb64c-ffa1-49ff-b4f5-edd21fc07008 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.160399] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 599.160399] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52200e8b-123e-ccb7-e48f-6bdac6601d48" [ 599.160399] env[65788]: _type = "Task" [ 599.160399] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.163484] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661681, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.172541] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52200e8b-123e-ccb7-e48f-6bdac6601d48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.218021] env[65788]: WARNING neutronclient.v2_0.client [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.219332] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.219332] env[65788]: WARNING openstack [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.253495] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 599.366764] env[65788]: WARNING neutronclient.v2_0.client [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.367838] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.367838] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.397156] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 599.459043] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.459337] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "5b701040-025c-4246-ad54-f2cf478e998d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.476024] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661680, 'name': CreateVM_Task, 'duration_secs': 0.508912} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.480153] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 599.481169] env[65788]: WARNING neutronclient.v2_0.client [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.482127] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.482127] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 599.482127] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 599.483730] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24c4367e-4233-422f-9b47-1cbbccdfbafb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.488905] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 599.488905] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52abe9d9-c66d-05f6-1c15-18aea6c4603f" [ 599.488905] env[65788]: _type = "Task" [ 599.488905] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.503680] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52abe9d9-c66d-05f6-1c15-18aea6c4603f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.514150] env[65788]: DEBUG nova.network.neutron [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Updating instance_info_cache with network_info: [{"id": "752afe50-54ab-4302-9e95-0147367eacf8", "address": "fa:16:3e:16:34:02", "network": {"id": "de01c168-c92a-493c-b22b-93f3e12c8a6c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-18798383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4bce3aed585840338789b5058627b1a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap752afe50-54", "ovs_interfaceid": "752afe50-54ab-4302-9e95-0147367eacf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 599.543934] env[65788]: DEBUG nova.compute.manager [req-85706137-5d21-4500-8f73-c84a22cec76d req-3fb164e9-68ef-4928-9771-920d789869dc service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Received event network-vif-plugged-752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 599.544101] env[65788]: DEBUG oslo_concurrency.lockutils [req-85706137-5d21-4500-8f73-c84a22cec76d req-3fb164e9-68ef-4928-9771-920d789869dc service nova] Acquiring lock "c06991cd-15b0-4d48-849a-140720ee07a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.545957] env[65788]: DEBUG oslo_concurrency.lockutils [req-85706137-5d21-4500-8f73-c84a22cec76d req-3fb164e9-68ef-4928-9771-920d789869dc service nova] Lock "c06991cd-15b0-4d48-849a-140720ee07a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.546400] env[65788]: DEBUG oslo_concurrency.lockutils [req-85706137-5d21-4500-8f73-c84a22cec76d req-3fb164e9-68ef-4928-9771-920d789869dc service nova] Lock "c06991cd-15b0-4d48-849a-140720ee07a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 599.546400] env[65788]: DEBUG nova.compute.manager [req-85706137-5d21-4500-8f73-c84a22cec76d req-3fb164e9-68ef-4928-9771-920d789869dc service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] No waiting events found dispatching network-vif-plugged-752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 599.546667] env[65788]: WARNING nova.compute.manager [req-85706137-5d21-4500-8f73-c84a22cec76d req-3fb164e9-68ef-4928-9771-920d789869dc service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Received unexpected event network-vif-plugged-752afe50-54ab-4302-9e95-0147367eacf8 for instance with vm_state building and task_state spawning. [ 599.602044] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540965} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.602280] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 39127362-8daf-43ff-99cb-ddb54c6fe095/39127362-8daf-43ff-99cb-ddb54c6fe095.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 599.603060] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 599.603060] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4500fbfc-6ff0-4cda-99fe-fe81343fba13 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.610494] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.610911] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.623406] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 599.623406] env[65788]: value = "task-4661682" [ 599.623406] env[65788]: _type = "Task" [ 599.623406] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.641288] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661682, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.666300] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661681, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51371} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.674301] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 48ac772b-3b3b-403e-814d-b5290b17fcad/48ac772b-3b3b-403e-814d-b5290b17fcad.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 599.678022] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 599.678022] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-981e80dc-776b-42b7-80a3-d63d4162fa09 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.684018] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52200e8b-123e-ccb7-e48f-6bdac6601d48, 'name': SearchDatastore_Task, 'duration_secs': 0.011909} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.686420] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 599.686420] env[65788]: value = "task-4661683" [ 599.686420] env[65788]: _type = "Task" [ 599.686420] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.686696] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21b2346b-7005-44c0-99de-8ae50466920a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.697209] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 599.697209] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52483cb2-2327-88d5-786b-b4a1140cee93" [ 599.697209] env[65788]: _type = "Task" [ 599.697209] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.700606] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.709768] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52483cb2-2327-88d5-786b-b4a1140cee93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.747840] env[65788]: WARNING neutronclient.v2_0.client [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 599.748535] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 599.748869] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 599.760089] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 599.760089] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.332s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 599.760089] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.692s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 599.760905] env[65788]: INFO nova.compute.claims [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.921434] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 599.954805] env[65788]: DEBUG nova.network.neutron [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Successfully updated port: a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 599.962343] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 599.985473] env[65788]: DEBUG nova.network.neutron [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Updated VIF entry in instance network info cache for port 4dd6a49e-86fd-442c-bd8d-4acf8587707f. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 599.985820] env[65788]: DEBUG nova.network.neutron [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Updating instance_info_cache with network_info: [{"id": "4dd6a49e-86fd-442c-bd8d-4acf8587707f", "address": "fa:16:3e:8a:b7:d2", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4dd6a49e-86", "ovs_interfaceid": "4dd6a49e-86fd-442c-bd8d-4acf8587707f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 600.005957] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52abe9d9-c66d-05f6-1c15-18aea6c4603f, 'name': SearchDatastore_Task, 'duration_secs': 0.05511} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.006330] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 600.006556] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 600.006781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.018336] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Releasing lock "refresh_cache-c06991cd-15b0-4d48-849a-140720ee07a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 600.018336] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Instance network_info: |[{"id": "752afe50-54ab-4302-9e95-0147367eacf8", "address": "fa:16:3e:16:34:02", "network": {"id": "de01c168-c92a-493c-b22b-93f3e12c8a6c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-18798383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4bce3aed585840338789b5058627b1a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap752afe50-54", "ovs_interfaceid": "752afe50-54ab-4302-9e95-0147367eacf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 600.018464] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:34:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '752afe50-54ab-4302-9e95-0147367eacf8', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 600.025455] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Creating folder: Project (4bce3aed585840338789b5058627b1a8). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.026022] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a341d28-9996-4ef7-accb-d467e1f05377 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.038868] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Created folder: Project (4bce3aed585840338789b5058627b1a8) in parent group-v910111. [ 600.040115] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Creating folder: Instances. Parent ref: group-v910124. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 600.040435] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f037f851-bc6c-4943-9e2b-30ac5a79caf8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.052569] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Created folder: Instances in parent group-v910124. [ 600.052900] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 600.053134] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 600.053359] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c15b1f2-7400-46f1-bcda-8b3277935a6d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.073366] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 600.073366] env[65788]: value = "task-4661686" [ 600.073366] env[65788]: _type = "Task" [ 600.073366] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.082382] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661686, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.135164] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661682, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078813} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.135782] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 600.137571] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28084794-eb9c-4807-b5ad-d50627ea5286 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.160316] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 39127362-8daf-43ff-99cb-ddb54c6fe095/39127362-8daf-43ff-99cb-ddb54c6fe095.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 600.160662] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df17986c-cecc-4959-b597-0459cf556dee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.181302] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 600.181302] env[65788]: value = "task-4661687" [ 600.181302] env[65788]: _type = "Task" [ 600.181302] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.191254] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661687, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.200407] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075173} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.201154] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 600.208209] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae249e27-6b7a-4372-9d0f-fd61a34ccc1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.216110] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52483cb2-2327-88d5-786b-b4a1140cee93, 'name': SearchDatastore_Task, 'duration_secs': 0.010869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.226227] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 600.226543] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] ed55713e-8bf1-4960-8309-47498b163d02/ed55713e-8bf1-4960-8309-47498b163d02.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 600.235317] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 48ac772b-3b3b-403e-814d-b5290b17fcad/48ac772b-3b3b-403e-814d-b5290b17fcad.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 600.235714] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.235911] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 600.236161] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c928ea40-249b-4033-8abe-98567d2326e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.238276] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19fd2552-6cc5-4d21-90b2-055a5c0269ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.258524] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25b77296-ba15-4ed7-b9e8-11e4ce145f9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.271104] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 600.271104] env[65788]: value = "task-4661689" [ 600.271104] env[65788]: _type = "Task" [ 600.271104] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.276789] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 600.276789] env[65788]: value = "task-4661688" [ 600.276789] env[65788]: _type = "Task" [ 600.276789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.285356] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 600.285751] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 600.294729] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e315778a-c3ac-4670-8e81-10b7ac191813 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.299059] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661688, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.299059] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661689, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.302463] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 600.302463] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52342fbb-370c-85cf-b89b-c8e1c73d483d" [ 600.302463] env[65788]: _type = "Task" [ 600.302463] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.317278] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52342fbb-370c-85cf-b89b-c8e1c73d483d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.457839] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.458194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquired lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.458306] env[65788]: DEBUG nova.network.neutron [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 600.493249] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Releasing lock "refresh_cache-48ac772b-3b3b-403e-814d-b5290b17fcad" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 600.493249] env[65788]: DEBUG nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Received event network-vif-plugged-ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 600.493249] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Acquiring lock "ed55713e-8bf1-4960-8309-47498b163d02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.493249] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Lock "ed55713e-8bf1-4960-8309-47498b163d02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 600.493249] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Lock "ed55713e-8bf1-4960-8309-47498b163d02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 600.493977] env[65788]: DEBUG nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] No waiting events found dispatching network-vif-plugged-ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 600.493977] env[65788]: WARNING nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Received unexpected event network-vif-plugged-ffa53d50-ad0e-438c-994a-b0b7a97f85f2 for instance with vm_state building and task_state spawning. [ 600.493977] env[65788]: DEBUG nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Received event network-changed-ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 600.493977] env[65788]: DEBUG nova.compute.manager [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Refreshing instance network info cache due to event network-changed-ffa53d50-ad0e-438c-994a-b0b7a97f85f2. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 600.493977] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Acquiring lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.494175] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Acquired lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.494175] env[65788]: DEBUG nova.network.neutron [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Refreshing network info cache for port ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 600.495198] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.586592] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661686, 'name': CreateVM_Task, 'duration_secs': 0.401242} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.586835] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 600.589347] env[65788]: WARNING neutronclient.v2_0.client [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 600.589347] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.589347] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 600.589347] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 600.589347] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a3aeb68-2b6a-469e-a495-cf90364a5c72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.597556] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 600.597556] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d5bdd0-cac6-d850-bba3-e45723cd43bc" [ 600.597556] env[65788]: _type = "Task" [ 600.597556] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.609654] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d5bdd0-cac6-d850-bba3-e45723cd43bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.697676] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661687, 'name': ReconfigVM_Task, 'duration_secs': 0.300128} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.698624] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 39127362-8daf-43ff-99cb-ddb54c6fe095/39127362-8daf-43ff-99cb-ddb54c6fe095.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 600.698876] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2277453-a553-409d-8191-0596759f4e0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.712170] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 600.712170] env[65788]: value = "task-4661690" [ 600.712170] env[65788]: _type = "Task" [ 600.712170] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.727100] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661690, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.801165] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661689, 'name': ReconfigVM_Task, 'duration_secs': 0.342947} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.801515] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661688, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.801629] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 48ac772b-3b3b-403e-814d-b5290b17fcad/48ac772b-3b3b-403e-814d-b5290b17fcad.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 600.802350] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b47c4860-fb95-4d6b-a72a-b9485e8246a1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.815604] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 600.815604] env[65788]: value = "task-4661691" [ 600.815604] env[65788]: _type = "Task" [ 600.815604] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.820805] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52342fbb-370c-85cf-b89b-c8e1c73d483d, 'name': SearchDatastore_Task, 'duration_secs': 0.015436} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.834393] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-693182a8-3924-4cb1-b007-52791e9ee3a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.837221] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 600.837488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 600.845970] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661691, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.847442] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 600.847442] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f933a6-1b75-ceeb-7366-5bf0d807c263" [ 600.847442] env[65788]: _type = "Task" [ 600.847442] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.865928] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f933a6-1b75-ceeb-7366-5bf0d807c263, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.962199] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.962407] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 600.999046] env[65788]: WARNING neutronclient.v2_0.client [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 600.999447] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 600.999822] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.007668] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f6145d0-8309-4b75-96e7-c7e358718e7d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.017295] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7a7536-52df-464b-acd2-75e703abea08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.050945] env[65788]: DEBUG nova.network.neutron [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 601.053363] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b0dfc1-3eeb-4269-b0a4-a311476fc843 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.063715] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b2ab45-b11c-45a1-ba93-fc798c9b88c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.086115] env[65788]: DEBUG nova.compute.provider_tree [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.108664] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d5bdd0-cac6-d850-bba3-e45723cd43bc, 'name': SearchDatastore_Task, 'duration_secs': 0.064188} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.108968] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.109212] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 601.109419] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.113073] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.113213] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.223012] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661690, 'name': Rename_Task, 'duration_secs': 0.187669} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.223373] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 601.223647] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2297cb93-7ace-43fc-a345-53199ad72e1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.231500] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 601.231500] env[65788]: value = "task-4661692" [ 601.231500] env[65788]: _type = "Task" [ 601.231500] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.241935] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.271124] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.271528] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.290198] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661688, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.297904] env[65788]: WARNING neutronclient.v2_0.client [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.298596] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.299023] env[65788]: WARNING openstack [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.330810] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661691, 'name': Rename_Task, 'duration_secs': 0.187713} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.330810] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 601.331045] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-617f0951-e96d-46bc-81ea-1d0de9320254 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.344256] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 601.344256] env[65788]: value = "task-4661693" [ 601.344256] env[65788]: _type = "Task" [ 601.344256] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.358821] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661693, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.362845] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f933a6-1b75-ceeb-7366-5bf0d807c263, 'name': SearchDatastore_Task, 'duration_secs': 0.022614} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.363273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.363545] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 98a2efec-cb20-4e95-9ed5-5b2057a48c39/98a2efec-cb20-4e95-9ed5-5b2057a48c39.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 601.363823] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.363997] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 601.364228] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b85810c-f2d1-4c9c-b6c4-f709ffb2945b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.366541] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f357da3b-90fd-42b4-8872-42979dd602ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.379992] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 601.379992] env[65788]: value = "task-4661694" [ 601.379992] env[65788]: _type = "Task" [ 601.379992] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.380309] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 601.380505] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 601.381406] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fb63ab0-ecfc-4993-8d4d-8be82bcc7ac1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.391353] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 601.391353] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5285d7ce-3811-156d-bdcd-6af9a102428d" [ 601.391353] env[65788]: _type = "Task" [ 601.391353] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.395106] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661694, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.404880] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5285d7ce-3811-156d-bdcd-6af9a102428d, 'name': SearchDatastore_Task, 'duration_secs': 0.009898} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.405707] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a841598-dbf2-4fd8-b676-c28c326cb2e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.413395] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 601.413395] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee441a-15aa-4b46-3f62-655aede763a8" [ 601.413395] env[65788]: _type = "Task" [ 601.413395] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.423725] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee441a-15aa-4b46-3f62-655aede763a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.469708] env[65788]: WARNING neutronclient.v2_0.client [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 601.469708] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 601.469708] env[65788]: WARNING openstack [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 601.522012] env[65788]: DEBUG nova.network.neutron [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Updating instance_info_cache with network_info: [{"id": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "address": "fa:16:3e:7f:d4:00", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0bbf2f7-74", "ovs_interfaceid": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 601.578309] env[65788]: DEBUG nova.compute.manager [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Received event network-vif-plugged-5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 601.578537] env[65788]: DEBUG oslo_concurrency.lockutils [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Acquiring lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 601.578588] env[65788]: DEBUG oslo_concurrency.lockutils [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 601.578969] env[65788]: DEBUG oslo_concurrency.lockutils [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 601.578969] env[65788]: DEBUG nova.compute.manager [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] No waiting events found dispatching network-vif-plugged-5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 601.579087] env[65788]: WARNING nova.compute.manager [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Received unexpected event network-vif-plugged-5cc1dfe1-7bda-4f6a-9828-254a5f93a508 for instance with vm_state building and task_state spawning. [ 601.579522] env[65788]: DEBUG nova.compute.manager [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Received event network-changed-5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 601.581537] env[65788]: DEBUG nova.compute.manager [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Refreshing instance network info cache due to event network-changed-5cc1dfe1-7bda-4f6a-9828-254a5f93a508. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 601.581537] env[65788]: DEBUG oslo_concurrency.lockutils [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Acquiring lock "refresh_cache-98a2efec-cb20-4e95-9ed5-5b2057a48c39" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.581537] env[65788]: DEBUG oslo_concurrency.lockutils [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Acquired lock "refresh_cache-98a2efec-cb20-4e95-9ed5-5b2057a48c39" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 601.581537] env[65788]: DEBUG nova.network.neutron [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Refreshing network info cache for port 5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 601.591630] env[65788]: DEBUG nova.scheduler.client.report [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 601.742451] env[65788]: DEBUG oslo_vmware.api [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661692, 'name': PowerOnVM_Task, 'duration_secs': 0.460296} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.742725] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 601.743193] env[65788]: INFO nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Took 10.13 seconds to spawn the instance on the hypervisor. [ 601.743453] env[65788]: DEBUG nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 601.744374] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bdb45d-75f1-458b-a2ff-757fd63cec72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.792158] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661688, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.856274] env[65788]: DEBUG oslo_vmware.api [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661693, 'name': PowerOnVM_Task, 'duration_secs': 0.513494} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.856620] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 601.856839] env[65788]: INFO nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Took 15.76 seconds to spawn the instance on the hypervisor. [ 601.857042] env[65788]: DEBUG nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 601.857999] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24ae5433-d2cc-4586-8e9b-b1cc28ab6d15 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.892204] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661694, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.925063] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee441a-15aa-4b46-3f62-655aede763a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009931} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.925349] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 601.925612] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] c06991cd-15b0-4d48-849a-140720ee07a0/c06991cd-15b0-4d48-849a-140720ee07a0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 601.925901] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7eda1b86-13af-4c97-898d-1ee48005625e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.934998] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 601.934998] env[65788]: value = "task-4661695" [ 601.934998] env[65788]: _type = "Task" [ 601.934998] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.944469] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661695, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.000124] env[65788]: DEBUG nova.network.neutron [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updated VIF entry in instance network info cache for port ffa53d50-ad0e-438c-994a-b0b7a97f85f2. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 602.000738] env[65788]: DEBUG nova.network.neutron [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [{"id": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "address": "fa:16:3e:2c:93:04", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa53d50-ad", "ovs_interfaceid": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 602.026167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Releasing lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 602.027390] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Instance network_info: |[{"id": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "address": "fa:16:3e:7f:d4:00", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0bbf2f7-74", "ovs_interfaceid": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 602.028436] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:d4:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0bbf2f7-7442-4457-a3e7-eba29d08da68', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 602.040579] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Creating folder: Project (becee551c9054443a78c0bc2765b3a17). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.041667] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c84b618-8aa3-4db0-aca5-ef1e956265b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.057000] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Created folder: Project (becee551c9054443a78c0bc2765b3a17) in parent group-v910111. [ 602.057000] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Creating folder: Instances. Parent ref: group-v910127. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 602.057726] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9bc63a22-e2e8-4fa4-82a5-381a63abc1e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.070226] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Created folder: Instances in parent group-v910127. [ 602.070483] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 602.070603] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 602.070825] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a6ab699-6ca1-4413-9f44-3734e006fb3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.091241] env[65788]: WARNING neutronclient.v2_0.client [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.091241] env[65788]: WARNING openstack [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.091241] env[65788]: WARNING openstack [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.099365] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 602.099913] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 602.103615] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.072s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 602.109058] env[65788]: INFO nova.compute.claims [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.114028] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 602.114028] env[65788]: value = "task-4661698" [ 602.114028] env[65788]: _type = "Task" [ 602.114028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.120406] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661698, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.263460] env[65788]: INFO nova.compute.manager [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Took 16.86 seconds to build instance. [ 602.296822] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661688, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.613841} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.297256] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] ed55713e-8bf1-4960-8309-47498b163d02/ed55713e-8bf1-4960-8309-47498b163d02.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 602.297915] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 602.298165] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30e18449-4bff-42c3-b809-4fff01a1f3ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.309289] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 602.309289] env[65788]: value = "task-4661699" [ 602.309289] env[65788]: _type = "Task" [ 602.309289] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.321457] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.383442] env[65788]: INFO nova.compute.manager [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Took 20.53 seconds to build instance. [ 602.396884] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661694, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.991147} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.397185] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 98a2efec-cb20-4e95-9ed5-5b2057a48c39/98a2efec-cb20-4e95-9ed5-5b2057a48c39.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 602.397409] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 602.397659] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc9a7122-d8cf-4eda-ae7d-ff06d7b12a76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.405807] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 602.405807] env[65788]: value = "task-4661700" [ 602.405807] env[65788]: _type = "Task" [ 602.405807] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.416723] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.454625] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661695, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.504460] env[65788]: DEBUG oslo_concurrency.lockutils [req-2c8e0e3e-30a6-4719-a816-ded81fcccfd3 req-539d3b19-fd7f-4238-ab4c-ebbbca4f5bf5 service nova] Releasing lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 602.562809] env[65788]: WARNING openstack [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.563313] env[65788]: WARNING openstack [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.608031] env[65788]: DEBUG nova.compute.utils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 602.609227] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 602.609878] env[65788]: DEBUG nova.network.neutron [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 602.611263] env[65788]: WARNING neutronclient.v2_0.client [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.611908] env[65788]: WARNING neutronclient.v2_0.client [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.612985] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 602.613552] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 602.640584] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661698, 'name': CreateVM_Task, 'duration_secs': 0.518037} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.640904] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 602.642726] env[65788]: WARNING neutronclient.v2_0.client [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 602.643207] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.643415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 602.643849] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 602.644808] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccffbe25-e6ee-4302-89ef-3e4f567ae1d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.653563] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 602.653563] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc5f00-9e36-168c-1141-df3f89fa666a" [ 602.653563] env[65788]: _type = "Task" [ 602.653563] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.665478] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc5f00-9e36-168c-1141-df3f89fa666a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.766205] env[65788]: DEBUG oslo_concurrency.lockutils [None req-038c39b9-c5ca-4eb9-b76b-f57472da3ad5 tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "39127362-8daf-43ff-99cb-ddb54c6fe095" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.372s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 602.824793] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083976} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.825518] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 602.826298] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c518a1e1-3d54-42b5-ad54-8626b251cca2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.853475] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] ed55713e-8bf1-4960-8309-47498b163d02/ed55713e-8bf1-4960-8309-47498b163d02.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 602.854297] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9b42db7-d2be-4fc9-8add-db209279da3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.878136] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 602.878136] env[65788]: value = "task-4661701" [ 602.878136] env[65788]: _type = "Task" [ 602.878136] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.888409] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661701, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.888863] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b5b7b3d-f65e-46b6-bf9e-d74db68cc840 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.052s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 602.891558] env[65788]: DEBUG nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Received event network-changed-752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 602.891783] env[65788]: DEBUG nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Refreshing instance network info cache due to event network-changed-752afe50-54ab-4302-9e95-0147367eacf8. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 602.892090] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Acquiring lock "refresh_cache-c06991cd-15b0-4d48-849a-140720ee07a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.892157] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Acquired lock "refresh_cache-c06991cd-15b0-4d48-849a-140720ee07a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 602.892287] env[65788]: DEBUG nova.network.neutron [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Refreshing network info cache for port 752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 602.921330] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119653} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.922042] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 602.922633] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0876c1-ef01-4ccb-b2f5-56bbde17117b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.948153] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] 98a2efec-cb20-4e95-9ed5-5b2057a48c39/98a2efec-cb20-4e95-9ed5-5b2057a48c39.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 602.948554] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d100237-c6cc-4e29-acba-96d314de4488 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.974205] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661695, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.832491} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.975461] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] c06991cd-15b0-4d48-849a-140720ee07a0/c06991cd-15b0-4d48-849a-140720ee07a0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 602.975678] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 602.975978] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 602.975978] env[65788]: value = "task-4661702" [ 602.975978] env[65788]: _type = "Task" [ 602.975978] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.976333] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56b4d831-0acf-46bd-b0b9-a1f1162f349f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.988868] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.991687] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 602.991687] env[65788]: value = "task-4661703" [ 602.991687] env[65788]: _type = "Task" [ 602.991687] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.123982] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 603.176074] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc5f00-9e36-168c-1141-df3f89fa666a, 'name': SearchDatastore_Task, 'duration_secs': 0.05434} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.176074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 603.176074] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 603.176074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.176277] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 603.176277] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 603.176277] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14f4f61b-760c-4428-8ae5-4827882a51a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.186345] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 603.186345] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 603.186345] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b3613f8-d2b7-425d-a607-086dbf7ebcaa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.192948] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 603.192948] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5206eeb2-a335-eed9-46e5-815298ee8a7a" [ 603.192948] env[65788]: _type = "Task" [ 603.192948] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.207718] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5206eeb2-a335-eed9-46e5-815298ee8a7a, 'name': SearchDatastore_Task, 'duration_secs': 0.009024} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.207718] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c775498-ad84-4ebc-a619-405621264b6e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.220322] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 603.220322] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523cb8e4-fa75-79cb-6ecd-c42777295f99" [ 603.220322] env[65788]: _type = "Task" [ 603.220322] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.229848] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523cb8e4-fa75-79cb-6ecd-c42777295f99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.269786] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 603.315437] env[65788]: DEBUG nova.policy [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b0c1ce04a704a3eaf0a0d4dd09f7e08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9674e2a5c86b48db8c865a50331ab846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 603.363028] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bd0478-f92a-41e5-a51d-93d0e1b9b9fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.371139] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abaffaa-0148-44a2-984c-828187c11350 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.414322] env[65788]: WARNING neutronclient.v2_0.client [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.415813] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.415813] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.424728] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f978ecb-67b0-4240-88d0-0428e0330a79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.433712] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661701, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.440048] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9745d7-0d8f-4d2a-993a-2d549e0d0c7a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.461957] env[65788]: DEBUG nova.compute.provider_tree [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.489136] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661702, 'name': ReconfigVM_Task, 'duration_secs': 0.297339} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.490135] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Reconfigured VM instance instance-00000004 to attach disk [datastore2] 98a2efec-cb20-4e95-9ed5-5b2057a48c39/98a2efec-cb20-4e95-9ed5-5b2057a48c39.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 603.490277] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b0f3b11-6f79-4674-b2d7-c5fab312ace5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.501657] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661703, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087102} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.503091] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 603.503498] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 603.503498] env[65788]: value = "task-4661704" [ 603.503498] env[65788]: _type = "Task" [ 603.503498] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.504319] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026d10ae-af7e-489e-8248-76f1af4889e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.532086] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] c06991cd-15b0-4d48-849a-140720ee07a0/c06991cd-15b0-4d48-849a-140720ee07a0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 603.536350] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0b4ff86-c392-4a14-b948-c57b97eab364 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.550663] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661704, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.551937] env[65788]: WARNING neutronclient.v2_0.client [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 603.552588] env[65788]: WARNING openstack [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 603.552938] env[65788]: WARNING openstack [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 603.568151] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 603.568151] env[65788]: value = "task-4661705" [ 603.568151] env[65788]: _type = "Task" [ 603.568151] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.577802] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661705, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.732705] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523cb8e4-fa75-79cb-6ecd-c42777295f99, 'name': SearchDatastore_Task, 'duration_secs': 0.012909} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.733027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 603.733229] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4bedb939-e86b-42bd-a490-a73086ecfd7d/4bedb939-e86b-42bd-a490-a73086ecfd7d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 603.733495] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29ea705a-3e9c-4632-bfd9-768f28ca4aba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.741637] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 603.741637] env[65788]: value = "task-4661706" [ 603.741637] env[65788]: _type = "Task" [ 603.741637] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.753283] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661706, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.801210] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 603.857292] env[65788]: DEBUG nova.network.neutron [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Successfully created port: d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 603.898966] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661701, 'name': ReconfigVM_Task, 'duration_secs': 0.839184} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.899472] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Reconfigured VM instance instance-00000002 to attach disk [datastore2] ed55713e-8bf1-4960-8309-47498b163d02/ed55713e-8bf1-4960-8309-47498b163d02.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 603.900284] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd9de175-83cb-44d9-9d25-715d652451af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.907958] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 603.907958] env[65788]: value = "task-4661707" [ 603.907958] env[65788]: _type = "Task" [ 603.907958] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.918426] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661707, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.965451] env[65788]: DEBUG nova.scheduler.client.report [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 604.023649] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661704, 'name': Rename_Task, 'duration_secs': 0.151886} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.024028] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.024330] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-38c53ef0-7dec-4305-a6c2-17aba2bb793e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.035118] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 604.035118] env[65788]: value = "task-4661708" [ 604.035118] env[65788]: _type = "Task" [ 604.035118] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.046246] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661708, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.085721] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661705, 'name': ReconfigVM_Task, 'duration_secs': 0.316379} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.086086] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Reconfigured VM instance instance-00000005 to attach disk [datastore2] c06991cd-15b0-4d48-849a-140720ee07a0/c06991cd-15b0-4d48-849a-140720ee07a0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 604.087185] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf730178-89e7-4e4b-b638-dfca545cfe5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.100047] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 604.100047] env[65788]: value = "task-4661709" [ 604.100047] env[65788]: _type = "Task" [ 604.100047] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.113491] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661709, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.134587] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 604.170758] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 604.170758] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 604.170758] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 604.171068] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 604.171068] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 604.171068] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 604.171291] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.171492] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 604.171664] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 604.171872] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 604.172544] env[65788]: DEBUG nova.virt.hardware [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 604.173332] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9649c7-3e19-4992-b398-0c5bd23c98ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.185966] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915cc4e7-1e3b-4d69-a6fb-f440db35d821 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.260596] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661706, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.420754] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661707, 'name': Rename_Task, 'duration_secs': 0.219817} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.421991] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.422305] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7eb07ef5-efc7-4660-a72a-c4336e6d5cde {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.433390] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 604.433390] env[65788]: value = "task-4661710" [ 604.433390] env[65788]: _type = "Task" [ 604.433390] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.444458] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.470714] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 604.471367] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 604.475876] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.554s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 604.477635] env[65788]: INFO nova.compute.claims [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.547971] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661708, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.610963] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661709, 'name': Rename_Task, 'duration_secs': 0.315226} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.611264] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 604.611729] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26ec7465-b802-44a3-9903-285cf61ab0c7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.625221] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 604.625221] env[65788]: value = "task-4661711" [ 604.625221] env[65788]: _type = "Task" [ 604.625221] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.636857] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.754216] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661706, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557048} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.754569] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4bedb939-e86b-42bd-a490-a73086ecfd7d/4bedb939-e86b-42bd-a490-a73086ecfd7d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 604.754831] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 604.755215] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-adb8e5cb-ba5d-4d27-986c-770b768a2ec5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.764036] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 604.764036] env[65788]: value = "task-4661712" [ 604.764036] env[65788]: _type = "Task" [ 604.764036] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.775197] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.838395] env[65788]: DEBUG nova.network.neutron [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Updated VIF entry in instance network info cache for port 5cc1dfe1-7bda-4f6a-9828-254a5f93a508. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 604.838727] env[65788]: DEBUG nova.network.neutron [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Updating instance_info_cache with network_info: [{"id": "5cc1dfe1-7bda-4f6a-9828-254a5f93a508", "address": "fa:16:3e:91:98:45", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.154", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5cc1dfe1-7b", "ovs_interfaceid": "5cc1dfe1-7bda-4f6a-9828-254a5f93a508", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 604.947859] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661710, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.983289] env[65788]: DEBUG nova.compute.utils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 604.988121] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 604.988345] env[65788]: DEBUG nova.network.neutron [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 604.988709] env[65788]: WARNING neutronclient.v2_0.client [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.991775] env[65788]: WARNING neutronclient.v2_0.client [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 604.992645] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 604.992826] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.050341] env[65788]: DEBUG oslo_vmware.api [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661708, 'name': PowerOnVM_Task, 'duration_secs': 0.758675} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.050689] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 605.050822] env[65788]: INFO nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Took 11.18 seconds to spawn the instance on the hypervisor. [ 605.051048] env[65788]: DEBUG nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 605.051933] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfaee319-3ecf-4436-935f-e021ccef87e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.139669] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661711, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.183513] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6ba90b-db9b-4ec1-82cf-93792a03f5a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.192191] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c669669b-7994-4897-a5d9-9e8f7916caf6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.224840] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5063231a-3633-4643-adfd-b2873c4208b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.234793] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75d8ce1a-166f-4b55-a5ed-d6228b65cb2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.249582] env[65788]: DEBUG nova.compute.provider_tree [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.274056] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070916} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.274354] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 605.275147] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb604e78-22a1-4e44-b80c-e41c15c882d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.300189] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] 4bedb939-e86b-42bd-a490-a73086ecfd7d/4bedb939-e86b-42bd-a490-a73086ecfd7d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 605.300898] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5821b7ab-aa34-4346-812d-e7892452ae9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.319948] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 605.320446] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 605.329161] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 605.329161] env[65788]: value = "task-4661713" [ 605.329161] env[65788]: _type = "Task" [ 605.329161] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.341474] env[65788]: DEBUG oslo_concurrency.lockutils [req-fc2050a2-ec7a-4034-b13e-7a4d81d44c87 req-79cc2f4c-ca3c-4674-b2d6-2deaa5417496 service nova] Releasing lock "refresh_cache-98a2efec-cb20-4e95-9ed5-5b2057a48c39" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 605.342152] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661713, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.444503] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661710, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.503325] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 605.572378] env[65788]: INFO nova.compute.manager [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Took 19.32 seconds to build instance. [ 605.642375] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661711, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.683865] env[65788]: DEBUG nova.policy [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c4a00121e214dc9a7b0866c8deee18e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffbe44d63c1d432e97849f15615329e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 605.752931] env[65788]: DEBUG nova.scheduler.client.report [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 605.841600] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661713, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.945600] env[65788]: DEBUG oslo_vmware.api [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661710, 'name': PowerOnVM_Task, 'duration_secs': 1.07164} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.945858] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 605.946086] env[65788]: INFO nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Took 16.57 seconds to spawn the instance on the hypervisor. [ 605.946267] env[65788]: DEBUG nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 605.947049] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf2ac32-9a1a-415c-8158-da9737d3e364 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.076991] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a1eb16f-62db-4775-9ca3-42a7204f6609 tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.837s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.119685] env[65788]: DEBUG nova.network.neutron [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Successfully updated port: d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 606.144433] env[65788]: DEBUG oslo_vmware.api [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661711, 'name': PowerOnVM_Task, 'duration_secs': 1.063574} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.145349] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 606.145973] env[65788]: INFO nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Took 9.95 seconds to spawn the instance on the hypervisor. [ 606.146536] env[65788]: DEBUG nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 606.150020] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850ad6f2-4d4a-4c7e-948b-98f036529981 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.262374] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.786s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 606.263308] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 606.267125] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.772s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 606.269991] env[65788]: INFO nova.compute.claims [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.342850] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661713, 'name': ReconfigVM_Task, 'duration_secs': 0.717209} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.343135] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Reconfigured VM instance instance-00000006 to attach disk [datastore2] 4bedb939-e86b-42bd-a490-a73086ecfd7d/4bedb939-e86b-42bd-a490-a73086ecfd7d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 606.343713] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f533960-1512-4a36-bd09-0655dee93c1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.351191] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 606.351191] env[65788]: value = "task-4661714" [ 606.351191] env[65788]: _type = "Task" [ 606.351191] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.359971] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661714, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.476693] env[65788]: INFO nova.compute.manager [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Took 21.44 seconds to build instance. [ 606.520823] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 606.555048] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 606.555312] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 606.555466] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 606.555645] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 606.555787] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 606.555928] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 606.556226] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.556396] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 606.556563] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 606.556726] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 606.556897] env[65788]: DEBUG nova.virt.hardware [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 606.557899] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b5357d-a998-4aee-9db6-d664c7718862 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.567544] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52d42da-0d69-42f2-b43a-d3a2916b3308 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.622821] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-141aefe8-1b95-4963-854d-da79ddf143f7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.624591] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-141aefe8-1b95-4963-854d-da79ddf143f7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 606.624591] env[65788]: DEBUG nova.network.neutron [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 606.670253] env[65788]: INFO nova.compute.manager [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Took 17.62 seconds to build instance. [ 606.776862] env[65788]: DEBUG nova.compute.utils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 606.780733] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 606.780958] env[65788]: DEBUG nova.network.neutron [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 606.781292] env[65788]: WARNING neutronclient.v2_0.client [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 606.781574] env[65788]: WARNING neutronclient.v2_0.client [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 606.782223] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 606.782625] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 606.862372] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661714, 'name': Rename_Task, 'duration_secs': 0.327221} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.862774] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 606.863124] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-393d58a0-91a0-47ee-8af3-e2376033d245 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.871671] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 606.871671] env[65788]: value = "task-4661715" [ 606.871671] env[65788]: _type = "Task" [ 606.871671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 606.881275] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661715, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 606.984223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9063bf26-fdd4-44b3-994a-7b6acac00eb4 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.962s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 607.126684] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.127184] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.175389] env[65788]: DEBUG oslo_concurrency.lockutils [None req-add46046-fd98-4562-b2b1-bb437499fe61 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "c06991cd-15b0-4d48-849a-140720ee07a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.149s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 607.247142] env[65788]: WARNING neutronclient.v2_0.client [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 607.247830] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 607.248187] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 607.281877] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 607.384292] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661715, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 607.503072] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144146ef-3fbd-448c-9d61-8ad8724e7a1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.515062] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b04d2ad-3ca4-44da-bdbb-90ea1010d78c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.553111] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54955ec4-c7ba-4f28-9fda-d29be5023df8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.564927] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab8915d-a433-4d44-b407-27d78c48488b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.583351] env[65788]: DEBUG nova.compute.provider_tree [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.785349] env[65788]: DEBUG nova.network.neutron [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Successfully created port: e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 607.838847] env[65788]: DEBUG nova.policy [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2625758e73c64384982cb820ea055cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b70d3c8627449eaa6372ebe3bd90233', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 607.845751] env[65788]: DEBUG nova.network.neutron [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 607.887016] env[65788]: DEBUG oslo_vmware.api [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661715, 'name': PowerOnVM_Task, 'duration_secs': 0.689162} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 607.889268] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 607.889481] env[65788]: INFO nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Took 9.42 seconds to spawn the instance on the hypervisor. [ 607.889686] env[65788]: DEBUG nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 607.891038] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212bafda-dae9-4448-9fe5-f84720a4d7e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.086515] env[65788]: DEBUG nova.scheduler.client.report [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 608.297204] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 608.331658] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 608.332224] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 608.332449] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 608.332521] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 608.332677] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 608.332824] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 608.333086] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.333253] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 608.333534] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 608.333793] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 608.334167] env[65788]: DEBUG nova.virt.hardware [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 608.335078] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e799f1b-156e-4162-b73a-5ef3f0a11d90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.344805] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd917b8-a323-4b29-891c-8025f440fc68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.411970] env[65788]: INFO nova.compute.manager [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Took 19.05 seconds to build instance. [ 608.593225] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 608.593725] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 608.599853] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.799s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.601427] env[65788]: INFO nova.compute.claims [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.865330] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 608.866110] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 608.908232] env[65788]: DEBUG nova.network.neutron [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Updated VIF entry in instance network info cache for port 752afe50-54ab-4302-9e95-0147367eacf8. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 608.908435] env[65788]: DEBUG nova.network.neutron [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Updating instance_info_cache with network_info: [{"id": "752afe50-54ab-4302-9e95-0147367eacf8", "address": "fa:16:3e:16:34:02", "network": {"id": "de01c168-c92a-493c-b22b-93f3e12c8a6c", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-18798383-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4bce3aed585840338789b5058627b1a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap752afe50-54", "ovs_interfaceid": "752afe50-54ab-4302-9e95-0147367eacf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 608.917778] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8aa0df7b-4b70-46d1-a6d7-555c9f93019d tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.566s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 609.101876] env[65788]: DEBUG nova.compute.utils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 609.103333] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 609.103606] env[65788]: DEBUG nova.network.neutron [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 609.104621] env[65788]: WARNING neutronclient.v2_0.client [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.105141] env[65788]: WARNING neutronclient.v2_0.client [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.106674] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.107051] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.413403] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Releasing lock "refresh_cache-c06991cd-15b0-4d48-849a-140720ee07a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 609.413689] env[65788]: DEBUG nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Received event network-vif-plugged-a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 609.414021] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Acquiring lock "4bedb939-e86b-42bd-a490-a73086ecfd7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.414297] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.414467] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 609.414634] env[65788]: DEBUG nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] No waiting events found dispatching network-vif-plugged-a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 609.414797] env[65788]: WARNING nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Received unexpected event network-vif-plugged-a0bbf2f7-7442-4457-a3e7-eba29d08da68 for instance with vm_state building and task_state spawning. [ 609.414951] env[65788]: DEBUG nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Received event network-changed-a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 609.415113] env[65788]: DEBUG nova.compute.manager [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Refreshing instance network info cache due to event network-changed-a0bbf2f7-7442-4457-a3e7-eba29d08da68. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 609.415292] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Acquiring lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.415421] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Acquired lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 609.415573] env[65788]: DEBUG nova.network.neutron [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Refreshing network info cache for port a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 609.612899] env[65788]: DEBUG nova.compute.manager [None req-54fa3797-dce9-4f1e-b001-099ac4669a68 tempest-ServerDiagnosticsV248Test-166683186 tempest-ServerDiagnosticsV248Test-166683186-project-admin] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 609.614430] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e92f305-a316-43a4-93bf-7efff27fded0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.618021] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 609.627201] env[65788]: INFO nova.compute.manager [None req-54fa3797-dce9-4f1e-b001-099ac4669a68 tempest-ServerDiagnosticsV248Test-166683186 tempest-ServerDiagnosticsV248Test-166683186-project-admin] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Retrieving diagnostics [ 609.628065] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95546b6a-93ec-4f16-8e27-e5e18ba5c7e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.768327] env[65788]: DEBUG nova.network.neutron [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Successfully created port: d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 609.851198] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e46ebe-71b7-4129-a62b-1d111d35d54f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.862628] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2233de-9921-4f1d-8df6-597735408cd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.902336] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7136cc42-658f-4620-ace7-1a6045fc2281 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.911866] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579542ca-a94e-4894-ba85-47cfaf4175b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.918421] env[65788]: WARNING neutronclient.v2_0.client [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 609.919333] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 609.919830] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 609.941856] env[65788]: DEBUG nova.compute.provider_tree [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.065402] env[65788]: DEBUG nova.network.neutron [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Successfully updated port: e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 610.107746] env[65788]: DEBUG nova.policy [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2c86cb0e3014d0099858da2a4a9cb51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4fa459531885434f9a977d68864b7042', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 610.445843] env[65788]: DEBUG nova.scheduler.client.report [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 610.486720] env[65788]: WARNING neutronclient.v2_0.client [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 610.487450] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 610.487881] env[65788]: WARNING openstack [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 610.571192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.571493] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 610.572023] env[65788]: DEBUG nova.network.neutron [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 610.630182] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 610.664594] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 610.665167] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 610.665465] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 610.665675] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 610.665814] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 610.665955] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 610.666180] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.666340] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 610.666498] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 610.666653] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 610.666823] env[65788]: DEBUG nova.virt.hardware [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 610.669547] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f10ff2f-2a46-4c1b-8d4f-4dfc4728a583 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.678670] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac78c6a-b320-48ef-bd5c-fada3743f128 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.815340] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Acquiring lock "48ac772b-3b3b-403e-814d-b5290b17fcad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.815609] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.815910] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Acquiring lock "48ac772b-3b3b-403e-814d-b5290b17fcad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.816279] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.816279] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 610.818871] env[65788]: INFO nova.compute.manager [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Terminating instance [ 610.952566] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 610.953184] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 611.076878] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.077460] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.328029] env[65788]: DEBUG nova.compute.manager [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 611.328029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.328029] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817be187-07a5-4ace-85ec-ade0fed2ca4d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.337058] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 611.337058] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49eae47c-cc48-4307-b5d3-a01f7a81067e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.346261] env[65788]: DEBUG oslo_vmware.api [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Waiting for the task: (returnval){ [ 611.346261] env[65788]: value = "task-4661716" [ 611.346261] env[65788]: _type = "Task" [ 611.346261] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.355475] env[65788]: DEBUG oslo_vmware.api [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Task: {'id': task-4661716, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.461293] env[65788]: DEBUG nova.compute.utils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 611.464057] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 611.464645] env[65788]: DEBUG nova.network.neutron [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 611.465211] env[65788]: WARNING neutronclient.v2_0.client [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.466741] env[65788]: WARNING neutronclient.v2_0.client [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.466955] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.467772] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.739478] env[65788]: DEBUG nova.network.neutron [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Successfully updated port: d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 611.856440] env[65788]: DEBUG oslo_vmware.api [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Task: {'id': task-4661716, 'name': PowerOffVM_Task, 'duration_secs': 0.469627} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.856959] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 611.857204] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 611.857505] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f51d02b0-c812-4646-8f89-24d1b5f1ed1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.929614] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 611.929882] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 611.930119] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Deleting the datastore file [datastore2] 48ac772b-3b3b-403e-814d-b5290b17fcad {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 611.931207] env[65788]: DEBUG nova.network.neutron [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Successfully created port: 709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 611.933621] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e452d54-6eca-41b4-a978-cbdb52f1bbc3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.942014] env[65788]: DEBUG oslo_vmware.api [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Waiting for the task: (returnval){ [ 611.942014] env[65788]: value = "task-4661718" [ 611.942014] env[65788]: _type = "Task" [ 611.942014] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.953044] env[65788]: DEBUG oslo_vmware.api [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Task: {'id': task-4661718, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.961517] env[65788]: DEBUG nova.network.neutron [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 611.977242] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 612.000795] env[65788]: DEBUG nova.policy [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd87bea227ec749ce944cacab8ac313ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9bd94f79bfc46e4948bb6758d3d74b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 612.016040] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.017434] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.027309] env[65788]: DEBUG nova.network.neutron [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Updating instance_info_cache with network_info: [{"id": "d564449f-c908-4d2b-9416-3ea9d3d371fc", "address": "fa:16:3e:2b:71:6d", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd564449f-c9", "ovs_interfaceid": "d564449f-c908-4d2b-9416-3ea9d3d371fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 612.243351] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.243497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 612.243682] env[65788]: DEBUG nova.network.neutron [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 612.266980] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.267460] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.453532] env[65788]: DEBUG oslo_vmware.api [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Task: {'id': task-4661718, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159133} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.453885] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 612.454354] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 612.454354] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.454511] env[65788]: INFO nova.compute.manager [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Took 1.13 seconds to destroy the instance on the hypervisor. [ 612.455194] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 612.455194] env[65788]: DEBUG nova.compute.manager [-] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 612.455194] env[65788]: DEBUG nova.network.neutron [-] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 612.455510] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 612.455841] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.456321] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.530198] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-141aefe8-1b95-4963-854d-da79ddf143f7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 612.530687] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance network_info: |[{"id": "d564449f-c908-4d2b-9416-3ea9d3d371fc", "address": "fa:16:3e:2b:71:6d", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd564449f-c9", "ovs_interfaceid": "d564449f-c908-4d2b-9416-3ea9d3d371fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 612.531465] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:71:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd564449f-c908-4d2b-9416-3ea9d3d371fc', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 612.542187] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating folder: Project (9674e2a5c86b48db8c865a50331ab846). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.543472] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b7ec560-98ab-40af-8d10-56e88463e52c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.560296] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created folder: Project (9674e2a5c86b48db8c865a50331ab846) in parent group-v910111. [ 612.560296] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating folder: Instances. Parent ref: group-v910130. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.560581] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-820b11c4-0303-4548-8fa9-128cddae21bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.574463] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created folder: Instances in parent group-v910130. [ 612.574871] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 612.575274] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 612.575920] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbc89d71-910a-463b-a7e7-2b2b545dc2a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.598859] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 612.598859] env[65788]: value = "task-4661721" [ 612.598859] env[65788]: _type = "Task" [ 612.598859] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.608359] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661721, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.749058] env[65788]: DEBUG nova.network.neutron [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Successfully created port: 855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 612.752663] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 612.753121] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 612.988118] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 613.017559] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 613.018540] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 613.018626] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 613.018805] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 613.018955] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 613.019242] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 613.019526] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.019762] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 613.019936] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 613.020271] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 613.020406] env[65788]: DEBUG nova.virt.hardware [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 613.021543] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad37a466-5337-4211-ad02-24b5dd24c285 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.032080] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07eb7cf-f6e7-49ff-980d-5aea057b4ff2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.091155] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.111213] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661721, 'name': CreateVM_Task} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.111396] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 613.111915] env[65788]: WARNING neutronclient.v2_0.client [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.112725] env[65788]: DEBUG oslo_vmware.service [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f3b439-4135-46ce-9acc-02647881a645 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.121154] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.121384] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.121802] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 613.122177] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51795a2a-e3a0-412a-853e-28c442d6814a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.125031] env[65788]: DEBUG nova.network.neutron [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 613.131164] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 613.131164] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525cc1bc-4db7-0477-6f31-2290e948432d" [ 613.131164] env[65788]: _type = "Task" [ 613.131164] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.140588] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525cc1bc-4db7-0477-6f31-2290e948432d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.563315] env[65788]: DEBUG nova.network.neutron [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Successfully updated port: 709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 613.595747] env[65788]: WARNING neutronclient.v2_0.client [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 613.596491] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.596870] env[65788]: WARNING openstack [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.644635] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 613.644635] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.644945] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.645919] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 613.645919] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.645919] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e5033e8-e827-4d7b-81cf-71acfd530635 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.665479] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.667080] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 613.668164] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b1cc36-d950-4774-92fc-28605b3bf806 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.677635] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d37043f-3654-4dbb-bf6d-7f6fb646c1cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.684555] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 613.684555] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524fee40-d6c4-4611-ce1c-b06944dd7908" [ 613.684555] env[65788]: _type = "Task" [ 613.684555] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.698406] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524fee40-d6c4-4611-ce1c-b06944dd7908, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.701509] env[65788]: DEBUG nova.compute.manager [req-a5279b13-0659-44c3-8df9-3d24f238ed1b req-1240b970-9ddf-40ff-a979-0b81ee66319b service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Received event network-vif-plugged-d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 613.701973] env[65788]: DEBUG oslo_concurrency.lockutils [req-a5279b13-0659-44c3-8df9-3d24f238ed1b req-1240b970-9ddf-40ff-a979-0b81ee66319b service nova] Acquiring lock "141aefe8-1b95-4963-854d-da79ddf143f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.702244] env[65788]: DEBUG oslo_concurrency.lockutils [req-a5279b13-0659-44c3-8df9-3d24f238ed1b req-1240b970-9ddf-40ff-a979-0b81ee66319b service nova] Lock "141aefe8-1b95-4963-854d-da79ddf143f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.702403] env[65788]: DEBUG oslo_concurrency.lockutils [req-a5279b13-0659-44c3-8df9-3d24f238ed1b req-1240b970-9ddf-40ff-a979-0b81ee66319b service nova] Lock "141aefe8-1b95-4963-854d-da79ddf143f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 613.702667] env[65788]: DEBUG nova.compute.manager [req-a5279b13-0659-44c3-8df9-3d24f238ed1b req-1240b970-9ddf-40ff-a979-0b81ee66319b service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] No waiting events found dispatching network-vif-plugged-d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 613.702744] env[65788]: WARNING nova.compute.manager [req-a5279b13-0659-44c3-8df9-3d24f238ed1b req-1240b970-9ddf-40ff-a979-0b81ee66319b service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Received unexpected event network-vif-plugged-d564449f-c908-4d2b-9416-3ea9d3d371fc for instance with vm_state building and task_state spawning. [ 614.067950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.068192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquired lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 614.068383] env[65788]: DEBUG nova.network.neutron [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 614.196277] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 614.196569] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 614.196823] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7442b52-f53b-4103-a5cb-97f0b85fa377 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.251095] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 614.253917] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Fetch image to [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 614.253917] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Downloading image file data d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk on the data store datastore1 {{(pid=65788) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 614.253917] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de9f131-b756-4da2-a987-dd3344192e38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.262886] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e2aea1-6d3e-4432-8482-82f83ccd1393 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.275222] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3084634d-7ef5-4ccd-8000-155bd37d3809 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.312445] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9261ab1-2dbd-4a0e-88fe-c37a43e724ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.320494] env[65788]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-23c9ce7f-4d69-42a4-8e7f-744224d510cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.357566] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Downloading image file data d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to the data store datastore1 {{(pid=65788) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 614.383106] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "350f3c86-063c-4ed5-895b-fc621b2fa825" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.383334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 614.417419] env[65788]: DEBUG oslo_vmware.rw_handles [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 614.496432] env[65788]: WARNING neutronclient.v2_0.client [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 614.497101] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.497449] env[65788]: WARNING openstack [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.571679] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.572226] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.662162] env[65788]: DEBUG nova.network.neutron [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Successfully updated port: 855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 614.687651] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 614.688429] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 614.888234] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 615.167190] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "refresh_cache-4d8f46cd-1c36-4e43-8110-66e9c991f28d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.167445] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquired lock "refresh_cache-4d8f46cd-1c36-4e43-8110-66e9c991f28d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 615.167665] env[65788]: DEBUG nova.network.neutron [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 615.197375] env[65788]: DEBUG nova.network.neutron [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 615.245552] env[65788]: DEBUG nova.network.neutron [-] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.277823] env[65788]: DEBUG nova.network.neutron [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Updated VIF entry in instance network info cache for port a0bbf2f7-7442-4457-a3e7-eba29d08da68. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 615.278318] env[65788]: DEBUG nova.network.neutron [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Updating instance_info_cache with network_info: [{"id": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "address": "fa:16:3e:7f:d4:00", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0bbf2f7-74", "ovs_interfaceid": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.319461] env[65788]: DEBUG oslo_vmware.rw_handles [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 615.319820] env[65788]: DEBUG oslo_vmware.rw_handles [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 615.416808] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Downloaded image file data d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk on the data store datastore1 {{(pid=65788) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 615.419414] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 615.419770] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk to [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 615.420081] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e195b1e-4af7-48b0-82fc-22a2ea6afe63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.424317] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 615.424588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.426260] env[65788]: INFO nova.compute.claims [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.434911] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 615.434911] env[65788]: value = "task-4661722" [ 615.434911] env[65788]: _type = "Task" [ 615.434911] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.452536] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661722, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.602468] env[65788]: DEBUG nova.network.neutron [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updating instance_info_cache with network_info: [{"id": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "address": "fa:16:3e:ff:eb:0c", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape936d0e7-de", "ovs_interfaceid": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 615.618314] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.620532] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.647200] env[65788]: WARNING neutronclient.v2_0.client [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.647200] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.648477] env[65788]: WARNING openstack [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.672022] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.672515] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.754047] env[65788]: INFO nova.compute.manager [-] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Took 3.30 seconds to deallocate network for instance. [ 615.781424] env[65788]: DEBUG oslo_concurrency.lockutils [req-d97dfae9-e1fc-4f34-bc03-ccbef381febd req-a66984bd-91ae-43a8-91f5-6ec306e04862 service nova] Releasing lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 615.962695] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661722, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.005925] env[65788]: DEBUG nova.network.neutron [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 616.106411] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 616.106814] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Instance network_info: |[{"id": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "address": "fa:16:3e:ff:eb:0c", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape936d0e7-de", "ovs_interfaceid": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 616.107417] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:eb:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e936d0e7-de51-4ac2-9b4a-08f999624c5c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 616.120364] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating folder: Project (ffbe44d63c1d432e97849f15615329e1). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 616.121283] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68282784-b21d-48c5-97f7-f04290985f5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.135642] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created folder: Project (ffbe44d63c1d432e97849f15615329e1) in parent group-v910111. [ 616.135642] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating folder: Instances. Parent ref: group-v910133. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 616.136311] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3cc4244-483f-428c-ac7d-63d96d8bfd16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.150743] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created folder: Instances in parent group-v910133. [ 616.151778] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 616.151778] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 616.151778] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e04cc4b0-5455-466f-bf88-a8ee73d59a79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.176272] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 616.176272] env[65788]: value = "task-4661725" [ 616.176272] env[65788]: _type = "Task" [ 616.176272] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.185552] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661725, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.263988] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.275544] env[65788]: WARNING neutronclient.v2_0.client [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.277479] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.277479] env[65788]: WARNING openstack [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.316261] env[65788]: DEBUG nova.network.neutron [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 616.331473] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.332039] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.450532] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661722, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.722133} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.451305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk to [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 616.451780] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/tmp-sparse.vmdk {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 616.452229] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16251ed6-35b9-407c-9907-861c3891c1af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.461425] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 616.461425] env[65788]: value = "task-4661726" [ 616.461425] env[65788]: _type = "Task" [ 616.461425] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.477602] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.641545] env[65788]: DEBUG nova.network.neutron [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updating instance_info_cache with network_info: [{"id": "709378e7-1b93-4d0c-9649-3fb90a38b091", "address": "fa:16:3e:8a:05:2d", "network": {"id": "a8e54c54-a388-46c7-b7bd-556b975e8abd", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-501645486-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fa459531885434f9a977d68864b7042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap709378e7-1b", "ovs_interfaceid": "709378e7-1b93-4d0c-9649-3fb90a38b091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 616.656436] env[65788]: WARNING neutronclient.v2_0.client [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 616.657871] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.658616] env[65788]: WARNING openstack [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.697055] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661725, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.734031] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197aa585-ceb6-429e-884b-2717a7aa50de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.747477] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e37ea1-4c60-4553-a3b2-7530cdcffcae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.787298] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4e570c-671c-4f22-8648-b81294213b65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.793358] env[65788]: DEBUG nova.network.neutron [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Updating instance_info_cache with network_info: [{"id": "855a1c9a-f74e-4cca-afa1-a8bb99f21513", "address": "fa:16:3e:c6:9c:40", "network": {"id": "b8a97270-da51-42c9-9e5e-10b2225f4bc2", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-661842959-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e9bd94f79bfc46e4948bb6758d3d74b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855a1c9a-f7", "ovs_interfaceid": "855a1c9a-f74e-4cca-afa1-a8bb99f21513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 616.799458] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bb5c94-c4f8-4978-8ff6-cf6898dccf58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.816379] env[65788]: DEBUG nova.compute.provider_tree [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.820020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 616.820020] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Instance network_info: |[{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 616.820318] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:57:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6e04a22-9f86-43c9-ac42-95191eaf454b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 616.828443] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating folder: Project (2b70d3c8627449eaa6372ebe3bd90233). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 616.829043] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75cb1876-d101-42ab-aa79-e91c6c0101cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.842199] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created folder: Project (2b70d3c8627449eaa6372ebe3bd90233) in parent group-v910111. [ 616.842423] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating folder: Instances. Parent ref: group-v910136. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 616.842724] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7479ea55-094f-43c7-b2d6-2222df9572fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.854323] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created folder: Instances in parent group-v910136. [ 616.854583] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 616.854796] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 616.855015] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7de86ca2-16eb-4880-9df6-faac3ae7b3bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.877144] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 616.877144] env[65788]: value = "task-4661729" [ 616.877144] env[65788]: _type = "Task" [ 616.877144] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.885981] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661729, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.972128] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027918} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.972407] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 616.972571] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Moving file from [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b to [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b. {{(pid=65788) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 616.972828] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-bb20abd6-f3ae-474e-abd2-93dc934d611a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.981577] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 616.981577] env[65788]: value = "task-4661730" [ 616.981577] env[65788]: _type = "Task" [ 616.981577] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.994586] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661730, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.146531] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Releasing lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.147434] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Instance network_info: |[{"id": "709378e7-1b93-4d0c-9649-3fb90a38b091", "address": "fa:16:3e:8a:05:2d", "network": {"id": "a8e54c54-a388-46c7-b7bd-556b975e8abd", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-501645486-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fa459531885434f9a977d68864b7042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap709378e7-1b", "ovs_interfaceid": "709378e7-1b93-4d0c-9649-3fb90a38b091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 617.149722] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:05:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '709378e7-1b93-4d0c-9649-3fb90a38b091', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 617.158197] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Creating folder: Project (4fa459531885434f9a977d68864b7042). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.158197] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98a2b8d8-e2be-4c00-adbb-88f2edba96ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.172900] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Created folder: Project (4fa459531885434f9a977d68864b7042) in parent group-v910111. [ 617.172900] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Creating folder: Instances. Parent ref: group-v910139. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.172900] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e02ce660-caa9-4bfd-a239-e893b364a26c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.185652] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Created folder: Instances in parent group-v910139. [ 617.185652] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 617.185924] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 617.186206] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9729374f-ff80-4894-aeaa-efab39c8fd24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.208979] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661725, 'name': CreateVM_Task, 'duration_secs': 0.743369} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.209735] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 617.211032] env[65788]: WARNING neutronclient.v2_0.client [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.211032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.211192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.211447] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 617.211795] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44d68951-a009-4df1-a575-77f5deb8a4b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.216134] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 617.216134] env[65788]: value = "task-4661733" [ 617.216134] env[65788]: _type = "Task" [ 617.216134] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.220722] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 617.220722] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fadc10-181f-9af3-fa26-3ab9485d7a6b" [ 617.220722] env[65788]: _type = "Task" [ 617.220722] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.227931] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661733, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.236115] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fadc10-181f-9af3-fa26-3ab9485d7a6b, 'name': SearchDatastore_Task, 'duration_secs': 0.013066} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.237047] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.237047] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 617.237047] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.295897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Releasing lock "refresh_cache-4d8f46cd-1c36-4e43-8110-66e9c991f28d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.295897] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Instance network_info: |[{"id": "855a1c9a-f74e-4cca-afa1-a8bb99f21513", "address": "fa:16:3e:c6:9c:40", "network": {"id": "b8a97270-da51-42c9-9e5e-10b2225f4bc2", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-661842959-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "e9bd94f79bfc46e4948bb6758d3d74b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855a1c9a-f7", "ovs_interfaceid": "855a1c9a-f74e-4cca-afa1-a8bb99f21513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 617.296899] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:9c:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2ede0e6-8d7a-4018-bb37-25bf388e9867', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '855a1c9a-f74e-4cca-afa1-a8bb99f21513', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 617.305344] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Creating folder: Project (e9bd94f79bfc46e4948bb6758d3d74b0). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.305699] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f69eb8b7-f387-49bd-9028-81ea270a020d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.318557] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Created folder: Project (e9bd94f79bfc46e4948bb6758d3d74b0) in parent group-v910111. [ 617.319122] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Creating folder: Instances. Parent ref: group-v910142. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 617.320097] env[65788]: DEBUG nova.scheduler.client.report [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.325045] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53f33af2-5845-4355-b056-387868c7e2fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.334703] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Created folder: Instances in parent group-v910142. [ 617.335830] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 617.335830] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 617.335830] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9de3a970-abfd-4e51-9df0-beff5f1e983b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.358222] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 617.358222] env[65788]: value = "task-4661736" [ 617.358222] env[65788]: _type = "Task" [ 617.358222] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.368570] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661736, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.388339] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661729, 'name': CreateVM_Task, 'duration_secs': 0.449746} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.388524] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 617.389305] env[65788]: WARNING neutronclient.v2_0.client [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.389652] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.389809] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.390153] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 617.390509] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b30a75fe-1e2c-48b3-8d6b-9a2d23e132d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.397762] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 617.397762] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee035f-3649-8838-dd28-acb7df440a37" [ 617.397762] env[65788]: _type = "Task" [ 617.397762] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.412783] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee035f-3649-8838-dd28-acb7df440a37, 'name': SearchDatastore_Task, 'duration_secs': 0.011304} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.413093] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 617.413326] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 617.413649] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.493220] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661730, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.033381} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.493220] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] File moved {{(pid=65788) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 617.493220] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Cleaning up location [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 617.493341] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore1] vmware_temp/90f48e6c-ccd1-4d7f-ab5d-0b53d5511cbd {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 617.493532] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05ea62e7-f74c-4f1c-9bc1-2154e5f0e776 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.501791] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 617.501791] env[65788]: value = "task-4661737" [ 617.501791] env[65788]: _type = "Task" [ 617.501791] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.512073] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.686345] env[65788]: DEBUG nova.compute.manager [req-86cf3779-7f32-4fc6-930d-f80eeca40b76 req-0a21f802-1edf-4edf-8528-c4d793fceccc service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Received event network-vif-plugged-e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 617.686632] env[65788]: DEBUG oslo_concurrency.lockutils [req-86cf3779-7f32-4fc6-930d-f80eeca40b76 req-0a21f802-1edf-4edf-8528-c4d793fceccc service nova] Acquiring lock "efe1048b-50e9-4add-910a-607a95759c7a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.686744] env[65788]: DEBUG oslo_concurrency.lockutils [req-86cf3779-7f32-4fc6-930d-f80eeca40b76 req-0a21f802-1edf-4edf-8528-c4d793fceccc service nova] Lock "efe1048b-50e9-4add-910a-607a95759c7a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.686926] env[65788]: DEBUG oslo_concurrency.lockutils [req-86cf3779-7f32-4fc6-930d-f80eeca40b76 req-0a21f802-1edf-4edf-8528-c4d793fceccc service nova] Lock "efe1048b-50e9-4add-910a-607a95759c7a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.687089] env[65788]: DEBUG nova.compute.manager [req-86cf3779-7f32-4fc6-930d-f80eeca40b76 req-0a21f802-1edf-4edf-8528-c4d793fceccc service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] No waiting events found dispatching network-vif-plugged-e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 617.687266] env[65788]: WARNING nova.compute.manager [req-86cf3779-7f32-4fc6-930d-f80eeca40b76 req-0a21f802-1edf-4edf-8528-c4d793fceccc service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Received unexpected event network-vif-plugged-e936d0e7-de51-4ac2-9b4a-08f999624c5c for instance with vm_state building and task_state spawning. [ 617.730747] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661733, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.761969] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "b5967423-89e1-44eb-b4de-22d09adb51a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.761969] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "b5967423-89e1-44eb-b4de-22d09adb51a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.829966] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.829966] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 617.835022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.569s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.835022] env[65788]: DEBUG nova.objects.instance [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lazy-loading 'resources' on Instance uuid 48ac772b-3b3b-403e-814d-b5290b17fcad {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 617.872545] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661736, 'name': CreateVM_Task, 'duration_secs': 0.385979} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.872545] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 617.873024] env[65788]: WARNING neutronclient.v2_0.client [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.873678] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.873678] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 617.873931] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 617.874768] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b32b947-7392-4875-9f02-1c85c10561a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.881657] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 617.881657] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243360d-cbed-6d9e-9a52-f7216246f25d" [ 617.881657] env[65788]: _type = "Task" [ 617.881657] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.893436] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243360d-cbed-6d9e-9a52-f7216246f25d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.013617] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034116} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.014044] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 618.015688] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d589717-dcd9-42df-81f9-aaf596fa94a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.023214] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 618.023214] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5235a2bb-b70f-ddb7-b381-aeb6349d2cf4" [ 618.023214] env[65788]: _type = "Task" [ 618.023214] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.041640] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5235a2bb-b70f-ddb7-b381-aeb6349d2cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.011087} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.041967] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.042340] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 618.042674] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.042882] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.043153] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8af15d8f-9d58-409e-8fb9-1557f3bd669f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.045549] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db6c7b40-a84c-487b-b28f-718825870d51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.056446] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 618.056446] env[65788]: value = "task-4661738" [ 618.056446] env[65788]: _type = "Task" [ 618.056446] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.057348] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.057453] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.061417] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05f45643-c1ba-400a-b743-0a6eb66ca3ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.075494] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 618.075494] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5250b961-c801-a6a1-9191-3f9cd254a5af" [ 618.075494] env[65788]: _type = "Task" [ 618.075494] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.075494] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.084829] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5250b961-c801-a6a1-9191-3f9cd254a5af, 'name': SearchDatastore_Task, 'duration_secs': 0.011168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.087744] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b07cf290-2cce-4771-bec7-d0f3ac51680c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.091429] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 618.091429] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fd5cb1-a5eb-a286-94c0-3135a102b1d0" [ 618.091429] env[65788]: _type = "Task" [ 618.091429] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.103122] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fd5cb1-a5eb-a286-94c0-3135a102b1d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.228665] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661733, 'name': CreateVM_Task, 'duration_secs': 0.962436} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.228878] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 618.229447] env[65788]: WARNING neutronclient.v2_0.client [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.229824] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.266554] env[65788]: DEBUG nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 618.337210] env[65788]: DEBUG nova.compute.utils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 618.339766] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 618.339901] env[65788]: DEBUG nova.network.neutron [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 618.340291] env[65788]: WARNING neutronclient.v2_0.client [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.340856] env[65788]: WARNING neutronclient.v2_0.client [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 618.341460] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.341790] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.400478] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243360d-cbed-6d9e-9a52-f7216246f25d, 'name': SearchDatastore_Task, 'duration_secs': 0.010849} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.405086] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.405734] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 618.405734] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.406118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.406705] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 618.406909] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abd3447e-623d-473e-976e-f9ae9dca1474 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.414252] env[65788]: DEBUG nova.policy [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '18b72b92f6ee46ec8fc27a57cebbd659', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65eb2dae008549259c476b0b6e6ecde0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 618.416993] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 618.416993] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521cc259-eaa7-ace5-4cf7-324018a66a1a" [ 618.416993] env[65788]: _type = "Task" [ 618.416993] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.427797] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521cc259-eaa7-ace5-4cf7-324018a66a1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.570771] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.573335] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 618.573846] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 618.574034] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e91f57e-100a-41b8-84f4-780b507350ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.582864] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 618.582864] env[65788]: value = "task-4661739" [ 618.582864] env[65788]: _type = "Task" [ 618.582864] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.596974] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661739, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.611074] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fd5cb1-a5eb-a286-94c0-3135a102b1d0, 'name': SearchDatastore_Task, 'duration_secs': 0.009848} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.611074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.611414] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] efe1048b-50e9-4add-910a-607a95759c7a/efe1048b-50e9-4add-910a-607a95759c7a.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 618.611732] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.613623] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.613623] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5aef4669-3d6b-4151-ac15-0604e9739e2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.615129] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8f33a7f-3f79-46f4-b0ee-e0533ffef7f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.628387] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 618.628387] env[65788]: value = "task-4661740" [ 618.628387] env[65788]: _type = "Task" [ 618.628387] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.630115] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.630678] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.634756] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6381ca0-e707-41f8-8052-57e223ef0547 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.645212] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 618.645212] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524ffa0b-a599-1340-0076-9523592f9f33" [ 618.645212] env[65788]: _type = "Task" [ 618.645212] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.649571] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661740, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.657509] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "77dcec96-50e5-4753-b3cb-c6aec377bce2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.657708] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 618.660965] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2e2c96-c194-4fcc-8268-1a0541621854 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.671669] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524ffa0b-a599-1340-0076-9523592f9f33, 'name': SearchDatastore_Task, 'duration_secs': 0.019984} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.674142] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73a0b5f1-b548-4679-9024-ed3fa6f4a909 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.677962] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f284d9c-4c6d-4e97-982e-c01cc3c60ee6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.686497] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 618.686497] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e710c4-834e-64dc-cda8-17547b282652" [ 618.686497] env[65788]: _type = "Task" [ 618.686497] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.717645] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fb21d0-a5ba-48f6-aea7-608fdfad3168 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.730853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbce82f-d1a6-4b45-a36e-f47f9b78dc9b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.735244] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e710c4-834e-64dc-cda8-17547b282652, 'name': SearchDatastore_Task, 'duration_secs': 0.011927} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.735826] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.736064] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa/eceafff8-7d28-4b9b-ade6-5736d5977efa.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 618.736710] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 618.736939] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 618.737189] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9671c96-6ecd-4dc7-ba43-d20d977c019f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.748386] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-603a9c27-4d74-47a7-8d95-28e3ffed7230 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.751601] env[65788]: DEBUG nova.compute.provider_tree [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.759989] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 618.759989] env[65788]: value = "task-4661741" [ 618.759989] env[65788]: _type = "Task" [ 618.759989] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.761337] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 618.761504] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 618.766910] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30a3cfdc-8535-4e99-be39-227d10c8af4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.778652] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 618.778652] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e8ef5f-55d3-3be4-2615-d9c94615d191" [ 618.778652] env[65788]: _type = "Task" [ 618.778652] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.779274] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661741, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.792163] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e8ef5f-55d3-3be4-2615-d9c94615d191, 'name': SearchDatastore_Task, 'duration_secs': 0.01321} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.792855] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfd6dc35-1e7d-4ff6-bc67-da1173d6620c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.801398] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 618.801398] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523de360-fa5a-c4a8-fea9-af42b7ea4aeb" [ 618.801398] env[65788]: _type = "Task" [ 618.801398] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.805673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 618.812408] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523de360-fa5a-c4a8-fea9-af42b7ea4aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.841031] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 618.896046] env[65788]: DEBUG nova.network.neutron [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Successfully created port: c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 618.933622] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521cc259-eaa7-ace5-4cf7-324018a66a1a, 'name': SearchDatastore_Task, 'duration_secs': 0.06282} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.933622] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 618.933622] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 618.933622] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.097729] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661739, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060619} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.098031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 619.098882] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1d5f36-85fc-471f-b734-a1c9b0e69819 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.124215] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 619.124645] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c5d00d7-9d75-40e4-8d01-a9e6d2d4e786 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.152717] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661740, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.157492] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 619.157492] env[65788]: value = "task-4661742" [ 619.157492] env[65788]: _type = "Task" [ 619.157492] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.166285] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 619.169873] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661742, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.256256] env[65788]: DEBUG nova.scheduler.client.report [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 619.277237] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661741, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.317381] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523de360-fa5a-c4a8-fea9-af42b7ea4aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.018309} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.317381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 619.317381] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 4d8f46cd-1c36-4e43-8110-66e9c991f28d/4d8f46cd-1c36-4e43-8110-66e9c991f28d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 619.317381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.317683] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 619.317683] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c41b6c1-e6ef-4085-8d0d-190448a80f80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.320387] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a66d2de5-cadc-452b-b83d-0f41979017c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.330911] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 619.330911] env[65788]: value = "task-4661743" [ 619.330911] env[65788]: _type = "Task" [ 619.330911] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.341023] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 619.341023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 619.343192] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b10222f6-0025-40bc-9c11-ff49a6f1f175 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.345987] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.356919] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 619.356919] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c0cb68-1324-583a-b8e3-43003fcd4448" [ 619.356919] env[65788]: _type = "Task" [ 619.356919] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.370763] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c0cb68-1324-583a-b8e3-43003fcd4448, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.411815] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "54b0945e-c9f5-4726-928e-de1d0ded5aae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.411815] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "54b0945e-c9f5-4726-928e-de1d0ded5aae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.652764] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661740, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581229} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.653188] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] efe1048b-50e9-4add-910a-607a95759c7a/efe1048b-50e9-4add-910a-607a95759c7a.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 619.653634] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 619.653953] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2b536a9-7400-4deb-9ad2-ad99a9ef8655 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.667616] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661742, 'name': ReconfigVM_Task, 'duration_secs': 0.402278} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.668448] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Reconfigured VM instance instance-00000007 to attach disk [datastore1] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 619.669318] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 619.669318] env[65788]: value = "task-4661744" [ 619.669318] env[65788]: _type = "Task" [ 619.669318] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.669595] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62b4aad3-b447-4457-8026-c7b88e500271 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.690697] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 619.690697] env[65788]: value = "task-4661745" [ 619.690697] env[65788]: _type = "Task" [ 619.690697] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.691635] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661744, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.701798] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661745, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.710971] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.763335] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.766109] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.961s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.768449] env[65788]: INFO nova.compute.claims [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.787349] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661741, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.790278] env[65788]: DEBUG nova.compute.manager [None req-9dd8a3b9-aada-44fe-b63b-37770de31b21 tempest-ServerDiagnosticsTest-1113711413 tempest-ServerDiagnosticsTest-1113711413-project-admin] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 619.791438] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46a0765-7a8b-4623-9c31-0d5212edbedc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.796574] env[65788]: INFO nova.scheduler.client.report [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Deleted allocations for instance 48ac772b-3b3b-403e-814d-b5290b17fcad [ 619.800273] env[65788]: INFO nova.compute.manager [None req-9dd8a3b9-aada-44fe-b63b-37770de31b21 tempest-ServerDiagnosticsTest-1113711413 tempest-ServerDiagnosticsTest-1113711413-project-admin] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Retrieving diagnostics [ 619.804976] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed4e87a-db34-4c74-9e2d-2890fdc2c0d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.849115] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.854073] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 619.872147] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c0cb68-1324-583a-b8e3-43003fcd4448, 'name': SearchDatastore_Task, 'duration_secs': 0.055622} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.873044] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff5a5d6e-78e8-4c2f-88d9-d3f230bcec32 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.882117] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 619.882117] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529bba2e-e3d9-4293-34ec-aaeef84bfc5a" [ 619.882117] env[65788]: _type = "Task" [ 619.882117] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.888287] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 619.888545] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 619.888695] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 619.888871] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 619.889020] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 619.889155] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 619.889351] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.889499] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 619.889683] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 619.889844] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 619.890015] env[65788]: DEBUG nova.virt.hardware [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 619.891190] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854dd9c0-5899-406c-9cf5-17d9766c1831 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.897341] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529bba2e-e3d9-4293-34ec-aaeef84bfc5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.903533] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b06d789-cd07-409b-a518-63ebd1fc8113 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.919752] env[65788]: DEBUG nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 619.962618] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Received event network-changed-d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 619.962811] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Refreshing instance network info cache due to event network-changed-d564449f-c908-4d2b-9416-3ea9d3d371fc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 619.963548] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Acquiring lock "refresh_cache-141aefe8-1b95-4963-854d-da79ddf143f7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.966260] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Acquired lock "refresh_cache-141aefe8-1b95-4963-854d-da79ddf143f7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.966260] env[65788]: DEBUG nova.network.neutron [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Refreshing network info cache for port d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 620.182262] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661744, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07837} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.182554] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 620.183877] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532314c4-41eb-4b2b-bd3d-395425bdb660 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.208307] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] efe1048b-50e9-4add-910a-607a95759c7a/efe1048b-50e9-4add-910a-607a95759c7a.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 620.210681] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e49680dd-6473-472b-a46c-c4e2d4fbba06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.232689] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661745, 'name': Rename_Task, 'duration_secs': 0.163759} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.233857] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 620.234202] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 620.234202] env[65788]: value = "task-4661746" [ 620.234202] env[65788]: _type = "Task" [ 620.234202] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.234445] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5534dbcb-d252-4914-a727-17c81bc86ce9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.247718] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661746, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.249579] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 620.249579] env[65788]: value = "task-4661747" [ 620.249579] env[65788]: _type = "Task" [ 620.249579] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.259444] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.284810] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661741, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.313590] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61821ec9-72cf-4fb1-b46e-5e39ef7b01ec tempest-DeleteServersAdminTestJSON-532503792 tempest-DeleteServersAdminTestJSON-532503792-project-admin] Lock "48ac772b-3b3b-403e-814d-b5290b17fcad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.497s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 620.350045] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.395865] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529bba2e-e3d9-4293-34ec-aaeef84bfc5a, 'name': SearchDatastore_Task, 'duration_secs': 0.025988} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.396327] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.396502] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 5b701040-025c-4246-ad54-f2cf478e998d/5b701040-025c-4246-ad54-f2cf478e998d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 620.396787] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f490837a-0563-49a1-966c-062a90318231 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.405157] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 620.405157] env[65788]: value = "task-4661748" [ 620.405157] env[65788]: _type = "Task" [ 620.405157] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.415214] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.441650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.469774] env[65788]: WARNING neutronclient.v2_0.client [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 620.470459] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.471062] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.754117] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661746, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.770024] env[65788]: DEBUG oslo_vmware.api [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661747, 'name': PowerOnVM_Task, 'duration_secs': 0.489327} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.770024] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 620.770024] env[65788]: INFO nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Took 16.63 seconds to spawn the instance on the hypervisor. [ 620.770024] env[65788]: DEBUG nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 620.770738] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4ecddd-89ea-486b-9b4a-c6fd626f0789 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.795129] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661741, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.930567} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.795455] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa/eceafff8-7d28-4b9b-ade6-5736d5977efa.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 620.795826] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 620.796402] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78556093-d95d-4fed-9c33-674d0574c0dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.807547] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 620.807547] env[65788]: value = "task-4661749" [ 620.807547] env[65788]: _type = "Task" [ 620.807547] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.824680] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661749, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.856225] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661743, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.922928] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661748, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.959142] env[65788]: DEBUG nova.network.neutron [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Successfully updated port: c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 621.050863] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8963e79-77b2-48e8-a6ba-a9f7337681b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.061161] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d35e58-dfe7-455c-8f0b-0af5f7e4009a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.094420] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1590e45-fcf6-47e4-9c05-d96209f4d9f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.105112] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e65e29-06f6-4a80-8b56-774f01cdcc96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.121310] env[65788]: DEBUG nova.compute.provider_tree [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.257646] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661746, 'name': ReconfigVM_Task, 'duration_secs': 0.79906} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.257924] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Reconfigured VM instance instance-00000008 to attach disk [datastore1] efe1048b-50e9-4add-910a-607a95759c7a/efe1048b-50e9-4add-910a-607a95759c7a.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 621.258639] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-165d35f4-d635-408c-8be5-164f7cc6ab80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.268837] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 621.268837] env[65788]: value = "task-4661750" [ 621.268837] env[65788]: _type = "Task" [ 621.268837] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.280730] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661750, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.302061] env[65788]: INFO nova.compute.manager [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Took 27.26 seconds to build instance. [ 621.320594] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661749, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093034} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.321708] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.322773] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce18d13d-3c19-48dd-b194-42aadc6cd145 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.352962] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa/eceafff8-7d28-4b9b-ade6-5736d5977efa.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.357597] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f5c413a-2d84-47fe-a26c-649a59ba2487 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.378745] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661743, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.047045} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.380461] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 4d8f46cd-1c36-4e43-8110-66e9c991f28d/4d8f46cd-1c36-4e43-8110-66e9c991f28d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.380692] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 621.381042] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 621.381042] env[65788]: value = "task-4661751" [ 621.381042] env[65788]: _type = "Task" [ 621.381042] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.381612] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e04738e-fc5c-4b50-b2c4-63489afc5a02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.392991] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661751, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.394583] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 621.394583] env[65788]: value = "task-4661752" [ 621.394583] env[65788]: _type = "Task" [ 621.394583] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.405963] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661752, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.417581] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.981493} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.417831] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 5b701040-025c-4246-ad54-f2cf478e998d/5b701040-025c-4246-ad54-f2cf478e998d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.417958] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 621.418294] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b63a6229-94e7-4799-843f-fca2194687f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.427067] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 621.427067] env[65788]: value = "task-4661753" [ 621.427067] env[65788]: _type = "Task" [ 621.427067] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.438125] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661753, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.462961] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "refresh_cache-350f3c86-063c-4ed5-895b-fc621b2fa825" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.463122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquired lock "refresh_cache-350f3c86-063c-4ed5-895b-fc621b2fa825" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.464222] env[65788]: DEBUG nova.network.neutron [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 621.624695] env[65788]: DEBUG nova.scheduler.client.report [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 621.757061] env[65788]: DEBUG nova.compute.manager [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 621.783838] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661750, 'name': Rename_Task, 'duration_secs': 0.187643} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.787740] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 621.787740] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19314959-d7c0-483d-b951-f93fad1d6fdc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.793635] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 621.793635] env[65788]: value = "task-4661754" [ 621.793635] env[65788]: _type = "Task" [ 621.793635] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.805602] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87d50ff6-e30b-4d5d-a224-5efae959aefd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "141aefe8-1b95-4963-854d-da79ddf143f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.776s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 621.806442] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.898733] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661751, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.908920] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661752, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077462} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.909337] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.910147] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35ea4f8-2ba8-496a-af11-2b4bec8f5f32 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.935798] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] 4d8f46cd-1c36-4e43-8110-66e9c991f28d/4d8f46cd-1c36-4e43-8110-66e9c991f28d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 621.936504] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbacfb4b-2ff4-4c6e-9de3-ac5e33164204 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.964231] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661753, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07749} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.965927] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 621.966405] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 621.966405] env[65788]: value = "task-4661755" [ 621.966405] env[65788]: _type = "Task" [ 621.966405] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.966880] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.967365] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.976494] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95251def-945d-4c0d-8cec-e0e9086c9cc3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.990926] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661755, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.010125] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 5b701040-025c-4246-ad54-f2cf478e998d/5b701040-025c-4246-ad54-f2cf478e998d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 622.010611] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9582841-5d40-4390-ac37-0cffba50ca3a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.028192] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.032492] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.045754] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 622.045754] env[65788]: value = "task-4661756" [ 622.045754] env[65788]: _type = "Task" [ 622.045754] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.056324] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661756, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.133394] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.133991] env[65788]: DEBUG nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 622.137239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.426s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.138957] env[65788]: INFO nova.compute.claims [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.187556] env[65788]: DEBUG nova.network.neutron [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 622.284838] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.305673] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661754, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.397477] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661751, 'name': ReconfigVM_Task, 'duration_secs': 0.766919} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.398262] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Reconfigured VM instance instance-00000009 to attach disk [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa/eceafff8-7d28-4b9b-ade6-5736d5977efa.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.399524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "c06991cd-15b0-4d48-849a-140720ee07a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.400286] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "c06991cd-15b0-4d48-849a-140720ee07a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.400749] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "c06991cd-15b0-4d48-849a-140720ee07a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.402028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "c06991cd-15b0-4d48-849a-140720ee07a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.402028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "c06991cd-15b0-4d48-849a-140720ee07a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 622.405609] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e9ed04e-17e4-45f0-9fea-e8ab98cf8260 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.406625] env[65788]: INFO nova.compute.manager [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Terminating instance [ 622.417155] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 622.417155] env[65788]: value = "task-4661757" [ 622.417155] env[65788]: _type = "Task" [ 622.417155] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.432202] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661757, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.496828] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661755, 'name': ReconfigVM_Task, 'duration_secs': 0.500361} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.498220] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Reconfigured VM instance instance-0000000b to attach disk [datastore1] 4d8f46cd-1c36-4e43-8110-66e9c991f28d/4d8f46cd-1c36-4e43-8110-66e9c991f28d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.498220] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5db7b3ad-47ff-4b1b-9cda-f561b06f4891 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.508262] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 622.508262] env[65788]: value = "task-4661758" [ 622.508262] env[65788]: _type = "Task" [ 622.508262] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.524248] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661758, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.557492] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661756, 'name': ReconfigVM_Task, 'duration_secs': 0.471525} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.557828] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 5b701040-025c-4246-ad54-f2cf478e998d/5b701040-025c-4246-ad54-f2cf478e998d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.558302] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-433014cb-d89f-44d3-b29f-8af9908b105a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.567227] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 622.567227] env[65788]: value = "task-4661759" [ 622.567227] env[65788]: _type = "Task" [ 622.567227] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.577799] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661759, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.643647] env[65788]: DEBUG nova.compute.utils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 622.651661] env[65788]: DEBUG nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 622.805302] env[65788]: DEBUG oslo_vmware.api [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661754, 'name': PowerOnVM_Task, 'duration_secs': 0.545301} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.805579] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 622.805621] env[65788]: INFO nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Took 16.28 seconds to spawn the instance on the hypervisor. [ 622.805792] env[65788]: DEBUG nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 622.806580] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fccd493-1edf-48b1-933f-a923d8e1d9d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.869902] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.870467] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.892409] env[65788]: WARNING neutronclient.v2_0.client [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 622.893214] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.895164] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.915993] env[65788]: DEBUG nova.compute.manager [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 622.916178] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.917204] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11854d34-90e8-40bb-a974-db52848c3efd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.933920] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661757, 'name': Rename_Task, 'duration_secs': 0.344265} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.938518] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 622.938518] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 622.938518] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0fb28e5-e00f-43b6-b2c9-d904f62d6878 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.940970] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0985e08e-1751-4def-b8bd-487c68def756 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.951024] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 622.951024] env[65788]: value = "task-4661761" [ 622.951024] env[65788]: _type = "Task" [ 622.951024] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.956407] env[65788]: DEBUG oslo_vmware.api [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 622.956407] env[65788]: value = "task-4661760" [ 622.956407] env[65788]: _type = "Task" [ 622.956407] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.965875] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.974415] env[65788]: DEBUG oslo_vmware.api [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.986184] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "5cef20ed-d9f0-4237-ae9d-da401b953904" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.986519] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.021654] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661758, 'name': Rename_Task, 'duration_secs': 0.161548} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.022314] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 623.022587] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d820860-5571-4ba5-af57-d8a7161f68cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.032973] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 623.032973] env[65788]: value = "task-4661762" [ 623.032973] env[65788]: _type = "Task" [ 623.032973] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.045535] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.079536] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661759, 'name': Rename_Task, 'duration_secs': 0.195386} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.079937] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 623.080243] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66522af6-f283-4776-bbf4-c472d27a4afa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.088218] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 623.088218] env[65788]: value = "task-4661763" [ 623.088218] env[65788]: _type = "Task" [ 623.088218] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.097303] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661763, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.156790] env[65788]: DEBUG nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 623.328665] env[65788]: INFO nova.compute.manager [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Took 25.33 seconds to build instance. [ 623.463291] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661761, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.472922] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4319b1-2359-4f20-b2a8-031085a6e53e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.484025] env[65788]: DEBUG oslo_vmware.api [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661760, 'name': PowerOffVM_Task, 'duration_secs': 0.272721} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.486187] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 623.486463] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 623.486796] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54fdfa5d-e414-4325-b548-8a9c68a6d61d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.489630] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8971252-cfd2-4a24-b6ff-bf06ee4f0e02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.493277] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 623.534311] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b1528a-ae01-435c-8640-1a55b000274c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.551552] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e9e39a-abc8-4a6d-922f-ccb7d2e1f75a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.564109] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661762, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.576566] env[65788]: DEBUG nova.compute.provider_tree [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.597986] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 623.598729] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 623.599500] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Deleting the datastore file [datastore2] c06991cd-15b0-4d48-849a-140720ee07a0 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 623.600458] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-231debe5-b372-47ba-80af-893c667f4740 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.607305] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661763, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.615414] env[65788]: DEBUG oslo_vmware.api [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for the task: (returnval){ [ 623.615414] env[65788]: value = "task-4661765" [ 623.615414] env[65788]: _type = "Task" [ 623.615414] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.627357] env[65788]: DEBUG oslo_vmware.api [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.831871] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ab3ab16e-7619-4ba6-8ccd-013f97240b56 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "efe1048b-50e9-4add-910a-607a95759c7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.839s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.959451] env[65788]: WARNING neutronclient.v2_0.client [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 623.959896] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 623.960246] env[65788]: WARNING openstack [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.972493] env[65788]: DEBUG oslo_vmware.api [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661761, 'name': PowerOnVM_Task, 'duration_secs': 0.68465} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.972661] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 623.972885] env[65788]: INFO nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 15.68 seconds to spawn the instance on the hypervisor. [ 623.973195] env[65788]: DEBUG nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 623.975385] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab21ffa-c510-4dea-81f3-512fda10884f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.014359] env[65788]: DEBUG nova.network.neutron [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Updated VIF entry in instance network info cache for port d564449f-c908-4d2b-9416-3ea9d3d371fc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 624.014558] env[65788]: DEBUG nova.network.neutron [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Updating instance_info_cache with network_info: [{"id": "d564449f-c908-4d2b-9416-3ea9d3d371fc", "address": "fa:16:3e:2b:71:6d", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd564449f-c9", "ovs_interfaceid": "d564449f-c908-4d2b-9416-3ea9d3d371fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 624.047453] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.057807] env[65788]: DEBUG oslo_vmware.api [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661762, 'name': PowerOnVM_Task, 'duration_secs': 0.578752} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.058065] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 624.058255] env[65788]: INFO nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Took 11.07 seconds to spawn the instance on the hypervisor. [ 624.058420] env[65788]: DEBUG nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 624.059228] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672d7741-49d2-4ccc-9ea6-2f57e9aa7d38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.082520] env[65788]: DEBUG nova.scheduler.client.report [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 624.102655] env[65788]: DEBUG oslo_vmware.api [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4661763, 'name': PowerOnVM_Task, 'duration_secs': 0.676443} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.102943] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 624.104158] env[65788]: INFO nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Took 13.47 seconds to spawn the instance on the hypervisor. [ 624.104158] env[65788]: DEBUG nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 624.104368] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f467a182-8f88-43de-89af-b43633bb69ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.128207] env[65788]: DEBUG nova.network.neutron [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Updating instance_info_cache with network_info: [{"id": "c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b", "address": "fa:16:3e:02:2a:ba", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4fe2c1f-0c", "ovs_interfaceid": "c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 624.136840] env[65788]: DEBUG oslo_vmware.api [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Task: {'id': task-4661765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.137102] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 624.137314] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 624.137506] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.137691] env[65788]: INFO nova.compute.manager [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Took 1.22 seconds to destroy the instance on the hypervisor. [ 624.137941] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 624.138146] env[65788]: DEBUG nova.compute.manager [-] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 624.138237] env[65788]: DEBUG nova.network.neutron [-] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 624.138478] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 624.138985] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.139243] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.172333] env[65788]: DEBUG nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 624.196406] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 624.197110] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 624.197437] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 624.198810] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 624.198810] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 624.198810] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 624.198810] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.198810] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 624.199542] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 624.199542] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 624.199677] env[65788]: DEBUG nova.virt.hardware [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 624.201188] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e2d3a5-0c55-40ca-b678-a4c33c153447 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.210758] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ad8ff3-7a7d-4fb5-83f9-f632b809d62c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.227130] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 624.234408] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Creating folder: Project (4f2bada899b74039930165dd21153376). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.235323] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d626918-ab1f-4fc0-b789-d2e02019b88e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.242275] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 624.248208] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Created folder: Project (4f2bada899b74039930165dd21153376) in parent group-v910111. [ 624.248531] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Creating folder: Instances. Parent ref: group-v910145. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.248791] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-027d71f2-d82b-4339-86f2-e600e98a9e43 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.260969] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Created folder: Instances in parent group-v910145. [ 624.261286] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 624.261473] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 624.261692] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3204ecd9-1685-4a5e-ae57-2bcede64429b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.279899] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 624.279899] env[65788]: value = "task-4661768" [ 624.279899] env[65788]: _type = "Task" [ 624.279899] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.289234] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661768, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.496423] env[65788]: INFO nova.compute.manager [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 24.59 seconds to build instance. [ 624.518417] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Releasing lock "refresh_cache-141aefe8-1b95-4963-854d-da79ddf143f7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.518694] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Received event network-vif-plugged-709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.518889] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.519156] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Lock "5b701040-025c-4246-ad54-f2cf478e998d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.519292] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Lock "5b701040-025c-4246-ad54-f2cf478e998d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.519450] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] No waiting events found dispatching network-vif-plugged-709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 624.519609] env[65788]: WARNING nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Received unexpected event network-vif-plugged-709378e7-1b93-4d0c-9649-3fb90a38b091 for instance with vm_state building and task_state spawning. [ 624.519794] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Received event network-changed-709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.519970] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Refreshing instance network info cache due to event network-changed-709378e7-1b93-4d0c-9649-3fb90a38b091. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 624.520185] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Acquiring lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.520321] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Acquired lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.520472] env[65788]: DEBUG nova.network.neutron [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Refreshing network info cache for port 709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 624.586027] env[65788]: INFO nova.compute.manager [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Took 20.81 seconds to build instance. [ 624.591157] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.591667] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 624.594478] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.153s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.596392] env[65788]: INFO nova.compute.claims [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.633019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Releasing lock "refresh_cache-350f3c86-063c-4ed5-895b-fc621b2fa825" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.633019] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Instance network_info: |[{"id": "c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b", "address": "fa:16:3e:02:2a:ba", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4fe2c1f-0c", "ovs_interfaceid": "c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 624.634645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:2a:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 624.647238] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Creating folder: Project (65eb2dae008549259c476b0b6e6ecde0). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.647952] env[65788]: INFO nova.compute.manager [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Took 24.18 seconds to build instance. [ 624.649632] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e7cc4e9-dd85-462b-aa10-4ff83c28eeae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.665026] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Created folder: Project (65eb2dae008549259c476b0b6e6ecde0) in parent group-v910111. [ 624.665026] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Creating folder: Instances. Parent ref: group-v910148. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 624.665026] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e737179a-ad54-4848-a234-00063959c84c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.680187] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Created folder: Instances in parent group-v910148. [ 624.684026] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 624.684026] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 624.684026] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8bde2e9-7bf6-4637-8b2b-1c76313b7b7a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.716117] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 624.716117] env[65788]: value = "task-4661771" [ 624.716117] env[65788]: _type = "Task" [ 624.716117] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.727505] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661771, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.792461] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661768, 'name': CreateVM_Task, 'duration_secs': 0.469154} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.792461] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 624.793094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.793534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.793630] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 624.793845] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-139894a0-9996-4f64-b3cc-bad7477ca064 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.800254] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 624.800254] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525214cb-050a-5545-dc56-13b58e30bbdf" [ 624.800254] env[65788]: _type = "Task" [ 624.800254] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.809559] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525214cb-050a-5545-dc56-13b58e30bbdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.957065] env[65788]: DEBUG nova.compute.manager [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Received event network-changed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 624.957375] env[65788]: DEBUG nova.compute.manager [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Refreshing instance network info cache due to event network-changed. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 624.957512] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Acquiring lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.957666] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Acquired lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.957786] env[65788]: DEBUG nova.network.neutron [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 625.003456] env[65788]: DEBUG oslo_concurrency.lockutils [None req-912013ab-6b52-44ee-8789-97c390bcc9dc tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.111s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.026762] env[65788]: WARNING neutronclient.v2_0.client [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.026762] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.026762] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.087829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5afd7e97-d0da-4d31-8744-a271a0f7cdea tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.250s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.102870] env[65788]: DEBUG nova.compute.utils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 625.110361] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 625.110524] env[65788]: DEBUG nova.network.neutron [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 625.112290] env[65788]: WARNING neutronclient.v2_0.client [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.112290] env[65788]: WARNING neutronclient.v2_0.client [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.112290] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.112487] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.142028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "15aba949-c04c-4021-add8-2e9d4e4771db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.142028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "15aba949-c04c-4021-add8-2e9d4e4771db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.153220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bcaf55d4-069a-4482-a549-65064b9b9168 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "5b701040-025c-4246-ad54-f2cf478e998d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.694s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.192929] env[65788]: DEBUG nova.policy [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48473ea517f64d3ba0051de1e9e378b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6dcc0ac928ee42b9a926c98464c4ec95', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 625.228832] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661771, 'name': CreateVM_Task, 'duration_secs': 0.40195} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.229039] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 625.230517] env[65788]: WARNING neutronclient.v2_0.client [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.230517] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.323260] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525214cb-050a-5545-dc56-13b58e30bbdf, 'name': SearchDatastore_Task, 'duration_secs': 0.012745} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.323587] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.323892] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.324019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.324160] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.324346] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 625.324673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.324987] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 625.326299] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c838182-1f85-4cf8-b0c7-e5d90f0900b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.328409] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f267f0b-ed32-457d-894a-ac1c2c3dcad4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.339092] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 625.339092] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f6a5fd-4a47-b7f4-493a-7d8bd26389aa" [ 625.339092] env[65788]: _type = "Task" [ 625.339092] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.347456] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 625.347734] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 625.349048] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2542297d-a3c8-4bd0-9b34-29c62402e0d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.355034] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f6a5fd-4a47-b7f4-493a-7d8bd26389aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.359309] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 625.359309] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52520a65-9181-126c-4506-a601b1a08ef4" [ 625.359309] env[65788]: _type = "Task" [ 625.359309] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.372838] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52520a65-9181-126c-4506-a601b1a08ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.462529] env[65788]: WARNING neutronclient.v2_0.client [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 625.463210] env[65788]: WARNING openstack [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.464930] env[65788]: WARNING openstack [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.546870] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.547350] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.555774] env[65788]: DEBUG nova.network.neutron [-] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 625.608827] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 625.654319] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 625.852197] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f6a5fd-4a47-b7f4-493a-7d8bd26389aa, 'name': SearchDatastore_Task, 'duration_secs': 0.01864} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.855867] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.856434] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 625.856434] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.870655] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52520a65-9181-126c-4506-a601b1a08ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.027019} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.874091] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4810f143-44ae-45ea-90fd-c727de165939 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.884022] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 625.884022] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52389100-0ad5-234d-3def-6036120da83b" [ 625.884022] env[65788]: _type = "Task" [ 625.884022] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.897405] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52389100-0ad5-234d-3def-6036120da83b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.955799] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c61fb85-5544-40b0-806d-ee57ba308f55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.969289] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01694444-abc9-4045-aab3-cb02620b6716 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.007528] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fab477-0883-42b9-a935-14d539ab25f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.017620] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f5cff9-9adc-4abc-809a-2a071396b994 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.037208] env[65788]: DEBUG nova.compute.provider_tree [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.042096] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Received event network-changed-e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 626.042264] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Refreshing instance network info cache due to event network-changed-e936d0e7-de51-4ac2-9b4a-08f999624c5c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 626.042664] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquiring lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.042664] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquired lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 626.042805] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Refreshing network info cache for port e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 626.058648] env[65788]: INFO nova.compute.manager [-] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Took 1.92 seconds to deallocate network for instance. [ 626.125403] env[65788]: DEBUG nova.network.neutron [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Successfully created port: ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 626.190700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.205036] env[65788]: DEBUG nova.compute.manager [None req-861d8008-e8da-4497-aebe-d1a45bc79390 tempest-ServerDiagnosticsV248Test-166683186 tempest-ServerDiagnosticsV248Test-166683186-project-admin] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 626.206985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476140a1-b2c2-4109-8dbc-896ac96faa8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.220070] env[65788]: INFO nova.compute.manager [None req-861d8008-e8da-4497-aebe-d1a45bc79390 tempest-ServerDiagnosticsV248Test-166683186 tempest-ServerDiagnosticsV248Test-166683186-project-admin] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Retrieving diagnostics [ 626.221621] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3cafff-18af-4fe4-a599-beef055da409 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.342777] env[65788]: WARNING neutronclient.v2_0.client [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.342777] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.342777] env[65788]: WARNING openstack [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.398616] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52389100-0ad5-234d-3def-6036120da83b, 'name': SearchDatastore_Task, 'duration_secs': 0.018749} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.399018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.399348] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b5967423-89e1-44eb-b4de-22d09adb51a8/b5967423-89e1-44eb-b4de-22d09adb51a8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 626.399782] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 626.400015] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 626.400285] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16d52bc5-f2f3-4eac-bd04-d8d80fbce418 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.404274] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa764627-696f-484b-a66b-a3bc17add948 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.415375] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 626.415375] env[65788]: value = "task-4661772" [ 626.415375] env[65788]: _type = "Task" [ 626.415375] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.417657] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 626.417894] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 626.424063] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b4c51ef-4057-4756-bc28-f9649f4e2e75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.432578] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 626.432578] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c26f2e-2027-0515-e518-5e8ec25276c6" [ 626.432578] env[65788]: _type = "Task" [ 626.432578] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.437399] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661772, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.451957] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c26f2e-2027-0515-e518-5e8ec25276c6, 'name': SearchDatastore_Task, 'duration_secs': 0.013579} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.451957] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3bc62e5-29f4-441a-be93-5fc277954aa2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.458170] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 626.458170] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52813d8e-edc9-a2e3-1b2f-9d868ab581c1" [ 626.458170] env[65788]: _type = "Task" [ 626.458170] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.467970] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52813d8e-edc9-a2e3-1b2f-9d868ab581c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.548026] env[65788]: WARNING neutronclient.v2_0.client [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.548026] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.548026] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.560681] env[65788]: DEBUG nova.scheduler.client.report [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 626.566579] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.631672] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 626.672662] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:46:06Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1043931115',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1901957895',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 626.672953] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 626.673137] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 626.673321] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 626.673464] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 626.673605] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 626.673811] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.673967] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 626.676063] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 626.676273] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 626.676462] env[65788]: DEBUG nova.virt.hardware [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 626.677538] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63d9cad-813c-4dac-87dd-c6648918b634 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.688379] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6907c08a-d452-4c46-9929-b427124924a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.818151] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.818412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.818575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.818734] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.818930] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.821955] env[65788]: INFO nova.compute.manager [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Terminating instance [ 626.929624] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661772, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.972617] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52813d8e-edc9-a2e3-1b2f-9d868ab581c1, 'name': SearchDatastore_Task, 'duration_secs': 0.013652} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.973091] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.973091] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 350f3c86-063c-4ed5-895b-fc621b2fa825/350f3c86-063c-4ed5-895b-fc621b2fa825.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 626.973362] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de3a4206-66ae-43ad-a180-aad859526832 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.984403] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 626.984403] env[65788]: value = "task-4661773" [ 626.984403] env[65788]: _type = "Task" [ 626.984403] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.995229] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661773, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.006146] env[65788]: WARNING openstack [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.006545] env[65788]: WARNING openstack [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.068689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 627.069252] env[65788]: DEBUG nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 627.073520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.789s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.330279] env[65788]: DEBUG nova.compute.manager [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 627.330545] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.331648] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e39aca1-ada9-41ab-b03f-c95b9c9f9e6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.341319] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 627.341710] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eab81e22-b774-4f96-8ad4-f7f0e2254d4c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.348953] env[65788]: DEBUG oslo_vmware.api [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 627.348953] env[65788]: value = "task-4661774" [ 627.348953] env[65788]: _type = "Task" [ 627.348953] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.359361] env[65788]: DEBUG oslo_vmware.api [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661774, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.432330] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661772, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640338} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.432805] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b5967423-89e1-44eb-b4de-22d09adb51a8/b5967423-89e1-44eb-b4de-22d09adb51a8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 627.433185] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 627.433570] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37891ebb-a358-4de5-8916-17d1ac276f3d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.443288] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 627.443288] env[65788]: value = "task-4661775" [ 627.443288] env[65788]: _type = "Task" [ 627.443288] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.454073] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.496482] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661773, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.577759] env[65788]: DEBUG nova.compute.utils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 627.580520] env[65788]: DEBUG nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 627.585847] env[65788]: INFO nova.compute.claims [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.752695] env[65788]: DEBUG nova.network.neutron [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updated VIF entry in instance network info cache for port 709378e7-1b93-4d0c-9649-3fb90a38b091. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 627.754257] env[65788]: DEBUG nova.network.neutron [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updating instance_info_cache with network_info: [{"id": "709378e7-1b93-4d0c-9649-3fb90a38b091", "address": "fa:16:3e:8a:05:2d", "network": {"id": "a8e54c54-a388-46c7-b7bd-556b975e8abd", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-501645486-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fa459531885434f9a977d68864b7042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap709378e7-1b", "ovs_interfaceid": "709378e7-1b93-4d0c-9649-3fb90a38b091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 627.862485] env[65788]: DEBUG oslo_vmware.api [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661774, 'name': PowerOffVM_Task, 'duration_secs': 0.499099} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.863405] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 627.863591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 627.864023] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49e86f74-55e0-47e6-9e6f-9f7fda1dc6d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.962447] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088111} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.962447] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 627.962447] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d657234b-76fc-4d23-8f46-a3324fff655c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.985884] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] b5967423-89e1-44eb-b4de-22d09adb51a8/b5967423-89e1-44eb-b4de-22d09adb51a8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 627.990572] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36fa2c84-18db-4a7e-bef4-1c010afe4ee9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.018541] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661773, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.916892} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.020764] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 350f3c86-063c-4ed5-895b-fc621b2fa825/350f3c86-063c-4ed5-895b-fc621b2fa825.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 628.020764] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 628.021983] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 628.021983] env[65788]: value = "task-4661777" [ 628.021983] env[65788]: _type = "Task" [ 628.021983] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.021983] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fe9079d-c30f-4811-92a8-47fa26f30c3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.036119] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661777, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.037727] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 628.037727] env[65788]: value = "task-4661778" [ 628.037727] env[65788]: _type = "Task" [ 628.037727] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.047472] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661778, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.065898] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 628.066276] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 628.066513] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Deleting the datastore file [datastore2] 98a2efec-cb20-4e95-9ed5-5b2057a48c39 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 628.066880] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8bdbde6-9c21-4429-8bc0-ed6d403e2ae3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.075244] env[65788]: DEBUG oslo_vmware.api [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for the task: (returnval){ [ 628.075244] env[65788]: value = "task-4661779" [ 628.075244] env[65788]: _type = "Task" [ 628.075244] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.086202] env[65788]: DEBUG oslo_vmware.api [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.091233] env[65788]: DEBUG nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 628.096968] env[65788]: INFO nova.compute.resource_tracker [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating resource usage from migration 29816798-534f-4aa4-bd9b-815c21495fbd [ 628.149973] env[65788]: WARNING neutronclient.v2_0.client [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.150472] env[65788]: WARNING openstack [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.151142] env[65788]: WARNING openstack [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.186272] env[65788]: DEBUG nova.network.neutron [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Successfully updated port: ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 628.214641] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.215229] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.260798] env[65788]: DEBUG oslo_concurrency.lockutils [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] Releasing lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.262279] env[65788]: DEBUG nova.compute.manager [req-febb90a0-d28f-4166-be9c-c581a587891a req-bc0a374e-1651-4d99-adcc-ae048ab7f4dd service nova] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Received event network-vif-deleted-4dd6a49e-86fd-442c-bd8d-4acf8587707f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 628.340256] env[65788]: DEBUG nova.compute.manager [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Received event network-vif-plugged-c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 628.340256] env[65788]: DEBUG oslo_concurrency.lockutils [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Acquiring lock "350f3c86-063c-4ed5-895b-fc621b2fa825-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 628.340256] env[65788]: DEBUG oslo_concurrency.lockutils [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 628.340256] env[65788]: DEBUG oslo_concurrency.lockutils [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.340256] env[65788]: DEBUG nova.compute.manager [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] No waiting events found dispatching network-vif-plugged-c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 628.340712] env[65788]: WARNING nova.compute.manager [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Received unexpected event network-vif-plugged-c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b for instance with vm_state building and task_state spawning. [ 628.340712] env[65788]: DEBUG nova.compute.manager [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Received event network-changed-c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 628.340712] env[65788]: DEBUG nova.compute.manager [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Refreshing instance network info cache due to event network-changed-c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 628.340712] env[65788]: DEBUG oslo_concurrency.lockutils [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Acquiring lock "refresh_cache-350f3c86-063c-4ed5-895b-fc621b2fa825" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.341013] env[65788]: DEBUG oslo_concurrency.lockutils [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Acquired lock "refresh_cache-350f3c86-063c-4ed5-895b-fc621b2fa825" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.341013] env[65788]: DEBUG nova.network.neutron [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Refreshing network info cache for port c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 628.395234] env[65788]: DEBUG nova.network.neutron [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Updating instance_info_cache with network_info: [{"id": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "address": "fa:16:3e:7f:d4:00", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0bbf2f7-74", "ovs_interfaceid": "a0bbf2f7-7442-4457-a3e7-eba29d08da68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 628.468620] env[65788]: WARNING neutronclient.v2_0.client [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.471038] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.471038] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.516753] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fd9857-24d5-4f4a-b14f-40636fab7959 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.528853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4491e81e-4bdb-4a83-9f42-6f920486fb91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.567416] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.572020] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597d404f-883c-4d5a-b1bf-782f0c1d3d55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.584537] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661778, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07692} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.585264] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 628.586561] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb5a151-1fd6-4191-a839-c87a3d0eaeae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.594165] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6f2046-3606-4b5b-8537-c4594b9cb0b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.597339] env[65788]: DEBUG oslo_vmware.api [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Task: {'id': task-4661779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.361463} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.600650] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 628.600814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 628.601040] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.601218] env[65788]: INFO nova.compute.manager [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Took 1.27 seconds to destroy the instance on the hypervisor. [ 628.601480] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 628.606635] env[65788]: DEBUG nova.compute.manager [-] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 628.606635] env[65788]: DEBUG nova.network.neutron [-] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 628.606868] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.607390] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.607632] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.633795] env[65788]: DEBUG nova.compute.provider_tree [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.644831] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] 350f3c86-063c-4ed5-895b-fc621b2fa825/350f3c86-063c-4ed5-895b-fc621b2fa825.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 628.645457] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1d18e73-0f2e-4195-bb11-53e7fcf1cd4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.670551] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 628.670551] env[65788]: value = "task-4661780" [ 628.670551] env[65788]: _type = "Task" [ 628.670551] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.681709] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661780, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.689202] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.689341] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.690090] env[65788]: DEBUG nova.network.neutron [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 628.777694] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updated VIF entry in instance network info cache for port e936d0e7-de51-4ac2-9b4a-08f999624c5c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 628.778063] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updating instance_info_cache with network_info: [{"id": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "address": "fa:16:3e:ff:eb:0c", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape936d0e7-de", "ovs_interfaceid": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 628.843661] env[65788]: WARNING neutronclient.v2_0.client [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.844143] env[65788]: WARNING openstack [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.844497] env[65788]: WARNING openstack [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.853134] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 628.904330] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb5987f-5d45-459f-9420-90939e190912 tempest-ServerExternalEventsTest-1200067894 tempest-ServerExternalEventsTest-1200067894-project] Releasing lock "refresh_cache-4bedb939-e86b-42bd-a490-a73086ecfd7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.037761] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661777, 'name': ReconfigVM_Task, 'duration_secs': 0.778305} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.040627] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Reconfigured VM instance instance-0000000d to attach disk [datastore2] b5967423-89e1-44eb-b4de-22d09adb51a8/b5967423-89e1-44eb-b4de-22d09adb51a8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.041627] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a50df8b-82af-479d-a592-0660cd67d91f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.051237] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 629.051237] env[65788]: value = "task-4661781" [ 629.051237] env[65788]: _type = "Task" [ 629.051237] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.063220] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661781, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.134887] env[65788]: DEBUG nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 629.146435] env[65788]: DEBUG nova.scheduler.client.report [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.168618] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 629.168878] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.169057] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 629.169293] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.169444] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 629.169586] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 629.169819] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.170347] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 629.170347] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 629.170519] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 629.170798] env[65788]: DEBUG nova.virt.hardware [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 629.171881] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550af624-c74c-45c8-ba1b-9b97232e01e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.185288] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661780, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.188893] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8192d4c-4a5a-4edd-92e2-66fde2ca6ce7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.194187] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.194580] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.214494] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 629.220091] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Creating folder: Project (7f0aa49fbf1746a39782c5b300475ca3). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 629.221235] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ac0f905-da54-471a-baaf-ced640943bfb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.234767] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Created folder: Project (7f0aa49fbf1746a39782c5b300475ca3) in parent group-v910111. [ 629.235473] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Creating folder: Instances. Parent ref: group-v910151. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 629.235473] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e89026a5-02b0-4fdd-b95e-de0b8b673d49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.246669] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Created folder: Instances in parent group-v910151. [ 629.246971] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 629.247155] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 629.247372] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ac2022f-c72f-43ce-a0e8-44cee799a9bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.266600] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 629.266600] env[65788]: value = "task-4661784" [ 629.266600] env[65788]: _type = "Task" [ 629.266600] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.275765] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661784, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.281489] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Releasing lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.282140] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Received event network-vif-plugged-d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 629.282349] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 629.282550] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.282704] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.282880] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] No waiting events found dispatching network-vif-plugged-d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 629.283060] env[65788]: WARNING nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Received unexpected event network-vif-plugged-d6e04a22-9f86-43c9-ac42-95191eaf454b for instance with vm_state active and task_state None. [ 629.283223] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Received event network-changed-d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 629.284120] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Refreshing instance network info cache due to event network-changed-d6e04a22-9f86-43c9-ac42-95191eaf454b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 629.284331] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.284512] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.284657] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Refreshing network info cache for port d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 629.459963] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 629.460211] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.537194] env[65788]: DEBUG nova.network.neutron [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 629.574499] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661781, 'name': Rename_Task, 'duration_secs': 0.349873} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.574871] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 629.575411] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec51dfc6-e48c-415c-881a-023f232b86c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.585161] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 629.585161] env[65788]: value = "task-4661785" [ 629.585161] env[65788]: _type = "Task" [ 629.585161] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.603747] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661785, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.651768] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.578s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.652561] env[65788]: INFO nova.compute.manager [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Migrating [ 629.652885] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.653128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "compute-rpcapi-router" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.654379] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.609s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.658094] env[65788]: INFO nova.compute.claims [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.661387] env[65788]: INFO nova.compute.rpcapi [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 629.662190] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "compute-rpcapi-router" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.691728] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661780, 'name': ReconfigVM_Task, 'duration_secs': 0.678221} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.691914] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Reconfigured VM instance instance-0000000c to attach disk [datastore2] 350f3c86-063c-4ed5-895b-fc621b2fa825/350f3c86-063c-4ed5-895b-fc621b2fa825.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 629.692546] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ff41d65-3ad2-4463-9925-e538d7ad77a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.702216] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 629.702216] env[65788]: value = "task-4661786" [ 629.702216] env[65788]: _type = "Task" [ 629.702216] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.720590] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661786, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.785418] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661784, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.788912] env[65788]: WARNING neutronclient.v2_0.client [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.789860] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.790449] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.822768] env[65788]: DEBUG nova.network.neutron [-] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 629.933128] env[65788]: WARNING openstack [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.933801] env[65788]: WARNING openstack [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.948214] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.948656] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.962926] env[65788]: DEBUG nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 630.097824] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661785, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.182482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.182482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.182482] env[65788]: DEBUG nova.network.neutron [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 630.214376] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661786, 'name': Rename_Task, 'duration_secs': 0.183978} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.214376] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 630.214376] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77b56c9f-ef1e-49c8-9d64-e3e7a8d22956 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.223628] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 630.223628] env[65788]: value = "task-4661787" [ 630.223628] env[65788]: _type = "Task" [ 630.223628] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.235103] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.290449] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661784, 'name': CreateVM_Task, 'duration_secs': 0.782715} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.292682] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 630.297798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.298407] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.299395] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 630.300367] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.300657] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 630.302140] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3b97375-add6-4e01-a872-0cae05b70cab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.311434] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 630.311434] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e33e7c-3b74-73cc-7cb7-aa0db1f506e0" [ 630.311434] env[65788]: _type = "Task" [ 630.311434] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.329495] env[65788]: INFO nova.compute.manager [-] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Took 1.72 seconds to deallocate network for instance. [ 630.329860] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e33e7c-3b74-73cc-7cb7-aa0db1f506e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.426111] env[65788]: WARNING neutronclient.v2_0.client [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.428030] env[65788]: WARNING openstack [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.428030] env[65788]: WARNING openstack [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.456465] env[65788]: WARNING neutronclient.v2_0.client [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.457184] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.457563] env[65788]: WARNING openstack [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.476771] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.477200] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.494300] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.597811] env[65788]: DEBUG oslo_vmware.api [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661785, 'name': PowerOnVM_Task, 'duration_secs': 0.671998} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.598292] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 630.598601] env[65788]: INFO nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Took 6.43 seconds to spawn the instance on the hypervisor. [ 630.598997] env[65788]: DEBUG nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 630.600413] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3922a3d6-b58e-4e45-b762-96a9b4fb0e5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.685108] env[65788]: WARNING neutronclient.v2_0.client [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 630.686876] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.687370] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.739683] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661787, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.809272] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 630.823402] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e33e7c-3b74-73cc-7cb7-aa0db1f506e0, 'name': SearchDatastore_Task, 'duration_secs': 0.024992} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.823784] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 630.824207] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 630.824546] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.824704] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.825475] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 630.828590] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d57279d-a522-415b-b036-19d9c475d9e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.836209] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.843309] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 630.843643] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 630.845083] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df5f1e13-f4cf-4e34-a159-2e9e834cd944 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.858030] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 630.858030] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f9424-2094-3161-d963-74523718790e" [ 630.858030] env[65788]: _type = "Task" [ 630.858030] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.871656] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f9424-2094-3161-d963-74523718790e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.001337] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3b9206-a184-4e4f-b43f-8573d47fdfc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.011521] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09c3cac-a9b9-4d19-a083-812467c5bc90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.051135] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac59374c-0dc3-4600-a7a5-0a83f369797c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.061540] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0849e318-4125-42bc-9a6e-fe42d073718a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.082665] env[65788]: DEBUG nova.compute.provider_tree [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.119851] env[65788]: INFO nova.compute.manager [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Took 12.35 seconds to build instance. [ 631.241450] env[65788]: DEBUG oslo_vmware.api [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661787, 'name': PowerOnVM_Task, 'duration_secs': 0.876007} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.241833] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 631.242109] env[65788]: INFO nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Took 11.39 seconds to spawn the instance on the hypervisor. [ 631.242341] env[65788]: DEBUG nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 631.246029] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25976ca0-9975-4401-82a6-08c1c30b2bf1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.330029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.369995] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f9424-2094-3161-d963-74523718790e, 'name': SearchDatastore_Task, 'duration_secs': 0.047563} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.370655] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4e37004-7eec-4260-b27a-39f18dd071fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.377346] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 631.377346] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204f866-c314-0227-8070-ee84687e58de" [ 631.377346] env[65788]: _type = "Task" [ 631.377346] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.389843] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204f866-c314-0227-8070-ee84687e58de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.460418] env[65788]: DEBUG nova.network.neutron [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Updated VIF entry in instance network info cache for port c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 631.460793] env[65788]: DEBUG nova.network.neutron [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Updating instance_info_cache with network_info: [{"id": "c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b", "address": "fa:16:3e:02:2a:ba", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.73", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4fe2c1f-0c", "ovs_interfaceid": "c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 631.500799] env[65788]: DEBUG nova.network.neutron [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Updating instance_info_cache with network_info: [{"id": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "address": "fa:16:3e:7e:b8:93", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca17fd35-e8", "ovs_interfaceid": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 631.528659] env[65788]: WARNING neutronclient.v2_0.client [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.529315] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.529704] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.590700] env[65788]: DEBUG nova.scheduler.client.report [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 631.622914] env[65788]: DEBUG oslo_concurrency.lockutils [None req-434ef496-850b-403f-a513-f71a17214603 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "b5967423-89e1-44eb-b4de-22d09adb51a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.861s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.715734] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.716142] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.765555] env[65788]: INFO nova.compute.manager [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Took 16.37 seconds to build instance. [ 631.797817] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updated VIF entry in instance network info cache for port d6e04a22-9f86-43c9-ac42-95191eaf454b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 631.798238] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 631.872676] env[65788]: WARNING neutronclient.v2_0.client [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 631.873159] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.873516] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.896902] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204f866-c314-0227-8070-ee84687e58de, 'name': SearchDatastore_Task, 'duration_secs': 0.03373} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.897226] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.897495] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 631.897764] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d080677e-8a46-46a4-a309-33631c83736f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.907079] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 631.907079] env[65788]: value = "task-4661788" [ 631.907079] env[65788]: _type = "Task" [ 631.907079] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.921281] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.967702] env[65788]: DEBUG oslo_concurrency.lockutils [req-92a3c6ef-b58e-4215-8b05-190ba1d6c996 req-45a43f84-aab8-44b6-abce-3be964b853f4 service nova] Releasing lock "refresh_cache-350f3c86-063c-4ed5-895b-fc621b2fa825" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.004251] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Releasing lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.004645] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Instance network_info: |[{"id": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "address": "fa:16:3e:7e:b8:93", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca17fd35-e8", "ovs_interfaceid": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 632.005341] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:b8:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca17fd35-e868-4f50-ac40-b6da3a12b015', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 632.014877] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Creating folder: Project (6dcc0ac928ee42b9a926c98464c4ec95). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.015674] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87f16b78-8849-4add-a1bc-000d21c72189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.030188] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Created folder: Project (6dcc0ac928ee42b9a926c98464c4ec95) in parent group-v910111. [ 632.030503] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Creating folder: Instances. Parent ref: group-v910154. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 632.030802] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-916376b5-5e3e-47f4-8d58-3985be0f592a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.045509] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Created folder: Instances in parent group-v910154. [ 632.046182] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 632.046486] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 632.046732] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96c7d10e-f785-4916-bfd9-c8946cdb84ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.074729] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 632.074729] env[65788]: value = "task-4661791" [ 632.074729] env[65788]: _type = "Task" [ 632.074729] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.091025] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661791, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.095117] env[65788]: DEBUG nova.network.neutron [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [{"id": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "address": "fa:16:3e:2c:93:04", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa53d50-ad", "ovs_interfaceid": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.102127] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.102127] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 632.103524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.914s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.105506] env[65788]: INFO nova.compute.claims [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.276668] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bed5556a-e90e-4b20-87bb-cfcead15d3c4 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.893s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.291100] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "4bedb939-e86b-42bd-a490-a73086ecfd7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.291497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.291828] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "4bedb939-e86b-42bd-a490-a73086ecfd7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.292069] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.292239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.294445] env[65788]: INFO nova.compute.manager [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Terminating instance [ 632.304260] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.304260] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Received event network-vif-plugged-855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 632.304260] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquiring lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.304260] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.304260] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.304711] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] No waiting events found dispatching network-vif-plugged-855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 632.304711] env[65788]: WARNING nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Received unexpected event network-vif-plugged-855a1c9a-f74e-4cca-afa1-a8bb99f21513 for instance with vm_state active and task_state None. [ 632.304711] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Received event network-changed-855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 632.304711] env[65788]: DEBUG nova.compute.manager [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Refreshing instance network info cache due to event network-changed-855a1c9a-f74e-4cca-afa1-a8bb99f21513. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 632.304711] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquiring lock "refresh_cache-4d8f46cd-1c36-4e43-8110-66e9c991f28d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.305020] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Acquired lock "refresh_cache-4d8f46cd-1c36-4e43-8110-66e9c991f28d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.305020] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Refreshing network info cache for port 855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 632.346955] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "957f14fc-23f1-4334-b672-b3a75398f716" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.346955] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "957f14fc-23f1-4334-b672-b3a75398f716" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.421942] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661788, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.587392] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661791, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.598429] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 632.615061] env[65788]: DEBUG nova.compute.utils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 632.619210] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 632.619665] env[65788]: DEBUG nova.network.neutron [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 632.620175] env[65788]: WARNING neutronclient.v2_0.client [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.620978] env[65788]: WARNING neutronclient.v2_0.client [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.621837] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.622837] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.637254] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 632.701446] env[65788]: DEBUG nova.policy [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c028d9c053054a2f916507b237d37e60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06b3dfec7c0144928b4a59dbd27569b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 632.799148] env[65788]: DEBUG nova.compute.manager [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 632.799414] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.800434] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f702eab-3208-484e-9d4b-a7b25dcf3a81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.811468] env[65788]: WARNING neutronclient.v2_0.client [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.811748] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.812059] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.820299] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 632.821363] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1707453-8b67-4828-a534-edeca39d6dd7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.834599] env[65788]: DEBUG oslo_vmware.api [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 632.834599] env[65788]: value = "task-4661792" [ 632.834599] env[65788]: _type = "Task" [ 632.834599] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.847211] env[65788]: DEBUG oslo_vmware.api [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.848867] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 632.931764] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640779} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.933036] env[65788]: DEBUG nova.compute.manager [req-0829605c-77a9-4c07-b7c5-cc9ed32fdc86 req-79d1a968-685e-42a1-aced-b01fd56a77da service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Received event network-vif-plugged-ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 632.933229] env[65788]: DEBUG oslo_concurrency.lockutils [req-0829605c-77a9-4c07-b7c5-cc9ed32fdc86 req-79d1a968-685e-42a1-aced-b01fd56a77da service nova] Acquiring lock "77dcec96-50e5-4753-b3cb-c6aec377bce2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.933426] env[65788]: DEBUG oslo_concurrency.lockutils [req-0829605c-77a9-4c07-b7c5-cc9ed32fdc86 req-79d1a968-685e-42a1-aced-b01fd56a77da service nova] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.933581] env[65788]: DEBUG oslo_concurrency.lockutils [req-0829605c-77a9-4c07-b7c5-cc9ed32fdc86 req-79d1a968-685e-42a1-aced-b01fd56a77da service nova] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.933732] env[65788]: DEBUG nova.compute.manager [req-0829605c-77a9-4c07-b7c5-cc9ed32fdc86 req-79d1a968-685e-42a1-aced-b01fd56a77da service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] No waiting events found dispatching network-vif-plugged-ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 632.933884] env[65788]: WARNING nova.compute.manager [req-0829605c-77a9-4c07-b7c5-cc9ed32fdc86 req-79d1a968-685e-42a1-aced-b01fd56a77da service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Received unexpected event network-vif-plugged-ca17fd35-e868-4f50-ac40-b6da3a12b015 for instance with vm_state building and task_state spawning. [ 632.934305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 632.934511] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.935105] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18f00e03-e17d-4b53-9c57-8511c1b9d757 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.944638] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 632.944638] env[65788]: value = "task-4661793" [ 632.944638] env[65788]: _type = "Task" [ 632.944638] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.956526] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661793, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.046855] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "27ce0779-2fa2-44d2-a424-072183c022a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.047263] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "27ce0779-2fa2-44d2-a424-072183c022a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.090520] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661791, 'name': CreateVM_Task, 'duration_secs': 0.770971} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.090520] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 633.091511] env[65788]: WARNING neutronclient.v2_0.client [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.091511] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.091511] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.092730] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 633.092730] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7e3d97c-134a-4aac-a198-ca5d71193d48 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.098873] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 633.098873] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c9f101-ac24-c861-f421-52cfb92688d5" [ 633.098873] env[65788]: _type = "Task" [ 633.098873] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.114503] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c9f101-ac24-c861-f421-52cfb92688d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.129039] env[65788]: DEBUG nova.network.neutron [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Successfully created port: 1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 633.349146] env[65788]: DEBUG oslo_vmware.api [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661792, 'name': PowerOffVM_Task, 'duration_secs': 0.497632} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.352430] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 633.353146] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 633.355735] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae6d8b84-d702-4438-a36c-50b2130b3203 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.373542] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.388064] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "39127362-8daf-43ff-99cb-ddb54c6fe095" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.388383] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "39127362-8daf-43ff-99cb-ddb54c6fe095" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.388597] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "39127362-8daf-43ff-99cb-ddb54c6fe095-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.388866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "39127362-8daf-43ff-99cb-ddb54c6fe095-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 633.389235] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "39127362-8daf-43ff-99cb-ddb54c6fe095-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.393915] env[65788]: INFO nova.compute.manager [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Terminating instance [ 633.450978] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 633.451440] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 633.452239] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Deleting the datastore file [datastore2] 4bedb939-e86b-42bd-a490-a73086ecfd7d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 633.453060] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33625f5b-4f22-4a45-bf9e-80fbf031e7ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.460714] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661793, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083186} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.464206] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 633.465763] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb04f574-43b6-46e5-9b87-ec15fa52c881 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.470029] env[65788]: DEBUG oslo_vmware.api [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for the task: (returnval){ [ 633.470029] env[65788]: value = "task-4661795" [ 633.470029] env[65788]: _type = "Task" [ 633.470029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.496652] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 633.497788] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-111a37db-846c-4e1d-a22b-25f43cb1a77e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.519603] env[65788]: DEBUG oslo_vmware.api [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661795, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.525844] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 633.525844] env[65788]: value = "task-4661796" [ 633.525844] env[65788]: _type = "Task" [ 633.525844] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.540700] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661796, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.549744] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 633.579525] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e94424-69db-4046-9cfc-1eb22caef248 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.588310] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccae2b5b-d223-43cd-9129-eac8e126f91e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.631787] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0c69d1-7b60-4300-bc3f-7bc2446bd713 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.645336] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e45bea8-f51e-4ffb-aabf-6673a4675f3d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.650195] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c9f101-ac24-c861-f421-52cfb92688d5, 'name': SearchDatastore_Task, 'duration_secs': 0.042176} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.651303] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 633.653688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.653929] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 633.654199] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.654352] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.654528] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 633.655706] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cf8fb7e9-a5a6-4192-9aff-224bf2e74b94 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.674322] env[65788]: DEBUG nova.compute.provider_tree [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.685647] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 633.685886] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 633.688034] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b0f0bb5-c3a4-4ae2-8d90-c6558ed1e17e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.692451] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 633.692728] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 633.692899] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 633.693106] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 633.693260] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 633.693403] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 633.693630] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.693787] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 633.694014] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 633.694141] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 633.694332] env[65788]: DEBUG nova.virt.hardware [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 633.696542] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290b3f08-1910-4859-bb5e-db316cff5a0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.702370] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 633.702370] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52355206-fc6c-cfc9-8803-cb1a12503c87" [ 633.702370] env[65788]: _type = "Task" [ 633.702370] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.708783] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d29a8e5-c975-426b-9be3-c287fab6ef65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.717687] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52355206-fc6c-cfc9-8803-cb1a12503c87, 'name': SearchDatastore_Task, 'duration_secs': 0.012328} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.719015] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4be6846-7bfa-4bb5-b88d-b62dc5396d28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.736042] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 633.736042] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b7fc04-9081-accd-635f-cde3031dfe31" [ 633.736042] env[65788]: _type = "Task" [ 633.736042] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.745781] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b7fc04-9081-accd-635f-cde3031dfe31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.843531] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.844051] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.898975] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "refresh_cache-39127362-8daf-43ff-99cb-ddb54c6fe095" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.899224] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquired lock "refresh_cache-39127362-8daf-43ff-99cb-ddb54c6fe095" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.899412] env[65788]: DEBUG nova.network.neutron [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 633.983257] env[65788]: DEBUG oslo_vmware.api [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Task: {'id': task-4661795, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342324} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.983603] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 633.983792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 633.983963] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.984149] env[65788]: INFO nova.compute.manager [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Took 1.18 seconds to destroy the instance on the hypervisor. [ 633.987438] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 633.987438] env[65788]: DEBUG nova.compute.manager [-] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 633.987438] env[65788]: DEBUG nova.network.neutron [-] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 633.987438] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 633.987438] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.987438] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.040399] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661796, 'name': ReconfigVM_Task, 'duration_secs': 0.44871} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.041435] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 634.041972] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bde1a52b-85a4-486d-a3fd-edec96eaa07b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.050621] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 634.050621] env[65788]: value = "task-4661797" [ 634.050621] env[65788]: _type = "Task" [ 634.050621] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.062273] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661797, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.083742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.139646] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8975363f-428d-479e-94ea-b87162cebf72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.162433] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 634.178587] env[65788]: DEBUG nova.scheduler.client.report [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 634.250565] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b7fc04-9081-accd-635f-cde3031dfe31, 'name': SearchDatastore_Task, 'duration_secs': 0.011903} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.252083] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.252678] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 77dcec96-50e5-4753-b3cb-c6aec377bce2/77dcec96-50e5-4753-b3cb-c6aec377bce2.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 634.253353] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e319ea5e-11d1-464d-b4bd-2d5e6630a453 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.268502] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 634.268502] env[65788]: value = "task-4661798" [ 634.268502] env[65788]: _type = "Task" [ 634.268502] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.280413] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.402410] env[65788]: WARNING neutronclient.v2_0.client [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.403244] env[65788]: WARNING openstack [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.403527] env[65788]: WARNING openstack [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.488894] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.527215] env[65788]: DEBUG nova.network.neutron [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 634.562195] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661797, 'name': Rename_Task, 'duration_secs': 0.312107} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.562493] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 634.562757] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1feddd07-cc63-4aac-b650-ee85c8294b1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.572965] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 634.572965] env[65788]: value = "task-4661799" [ 634.572965] env[65788]: _type = "Task" [ 634.572965] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.588688] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661799, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.592470] env[65788]: WARNING neutronclient.v2_0.client [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 634.594722] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 634.595154] env[65788]: WARNING openstack [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 634.671712] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 634.672037] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba276f78-8b69-490b-b94c-b8d843c2f7e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.682384] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 634.682384] env[65788]: value = "task-4661800" [ 634.682384] env[65788]: _type = "Task" [ 634.682384] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.691532] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.692840] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 634.694588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.128s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.695222] env[65788]: DEBUG nova.objects.instance [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lazy-loading 'resources' on Instance uuid c06991cd-15b0-4d48-849a-140720ee07a0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 634.706286] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.784183] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661798, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.928373] env[65788]: DEBUG nova.network.neutron [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Successfully updated port: 1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 634.980722] env[65788]: DEBUG nova.network.neutron [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 634.988255] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Updated VIF entry in instance network info cache for port 855a1c9a-f74e-4cca-afa1-a8bb99f21513. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 634.988611] env[65788]: DEBUG nova.network.neutron [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Updating instance_info_cache with network_info: [{"id": "855a1c9a-f74e-4cca-afa1-a8bb99f21513", "address": "fa:16:3e:c6:9c:40", "network": {"id": "b8a97270-da51-42c9-9e5e-10b2225f4bc2", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-661842959-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9bd94f79bfc46e4948bb6758d3d74b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap855a1c9a-f7", "ovs_interfaceid": "855a1c9a-f74e-4cca-afa1-a8bb99f21513", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 634.995889] env[65788]: DEBUG nova.compute.manager [req-c592db39-e35b-4787-b3e6-7471603a7cad req-6dc4e154-d4f3-4d0a-a50f-8be7e3dcb45b service nova] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Received event network-vif-deleted-752afe50-54ab-4302-9e95-0147367eacf8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 635.088868] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661799, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.199457] env[65788]: DEBUG nova.compute.utils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 635.203925] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661800, 'name': PowerOffVM_Task, 'duration_secs': 0.347605} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.203925] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 635.204186] env[65788]: DEBUG nova.network.neutron [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 635.205260] env[65788]: WARNING neutronclient.v2_0.client [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.205260] env[65788]: WARNING neutronclient.v2_0.client [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 635.205260] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.205635] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.213836] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 635.213836] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 635.285398] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.88012} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.285665] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 77dcec96-50e5-4753-b3cb-c6aec377bce2/77dcec96-50e5-4753-b3cb-c6aec377bce2.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.285881] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 635.289267] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2fc1b94d-61a9-478e-b68a-8f0fbf47afa3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.298181] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 635.298181] env[65788]: value = "task-4661801" [ 635.298181] env[65788]: _type = "Task" [ 635.298181] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.316084] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.363152] env[65788]: DEBUG nova.network.neutron [-] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 635.368798] env[65788]: DEBUG nova.policy [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c41f9b640034d60903ba8c0a3527d44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dc8fe6cf1ed4cdb91bf077edabce4b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 635.383091] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.383576] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.435189] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "refresh_cache-5cef20ed-d9f0-4237-ae9d-da401b953904" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.435857] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "refresh_cache-5cef20ed-d9f0-4237-ae9d-da401b953904" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.436177] env[65788]: DEBUG nova.network.neutron [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 635.493901] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Releasing lock "refresh_cache-39127362-8daf-43ff-99cb-ddb54c6fe095" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.494594] env[65788]: DEBUG nova.compute.manager [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 635.496639] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 635.496639] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441ab2e5-1409-4aba-83f7-b79d286af087 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.500976] env[65788]: DEBUG oslo_concurrency.lockutils [req-309d0d57-e182-40ce-a9ff-f69e49d9481c req-2a390df4-7183-4ef2-ad4e-9827d5682971 service nova] Releasing lock "refresh_cache-4d8f46cd-1c36-4e43-8110-66e9c991f28d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.509944] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 635.510289] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22ba15a6-c70a-488c-b41d-12a6da44e1fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.523598] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 635.523598] env[65788]: value = "task-4661802" [ 635.523598] env[65788]: _type = "Task" [ 635.523598] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.538424] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.588640] env[65788]: DEBUG oslo_vmware.api [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661799, 'name': PowerOnVM_Task, 'duration_secs': 0.635237} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.588925] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 635.589141] env[65788]: INFO nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Took 6.45 seconds to spawn the instance on the hypervisor. [ 635.590768] env[65788]: DEBUG nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 635.590768] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603af3c9-f727-4d02-87ce-3147da30cb7d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.601684] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1077daa-f9b5-4e26-bca4-0ceeda2b9e27 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.614911] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d2ea36-61ea-46a3-9e53-271b9ffda85d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.660574] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068928b1-6998-4bca-8e75-c89bd12105e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.672579] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad82b9d-c67a-4902-a80b-dd7b94f54d81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.689061] env[65788]: DEBUG nova.compute.provider_tree [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.718423] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 635.726033] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 635.726033] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 635.726033] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 635.726033] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 635.726573] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 635.726573] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 635.726573] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.726967] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 635.727324] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 635.727625] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 635.727990] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 635.735127] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20e2377f-2ca8-41f6-b7f1-e420719ad584 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.759741] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 635.759741] env[65788]: value = "task-4661803" [ 635.759741] env[65788]: _type = "Task" [ 635.759741] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.772060] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661803, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.810040] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075261} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.810486] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 635.811094] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e252dc1c-9311-46f1-9339-472590c960a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.843764] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 77dcec96-50e5-4753-b3cb-c6aec377bce2/77dcec96-50e5-4753-b3cb-c6aec377bce2.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 635.845508] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2ea63dc-eb9f-49ab-bab8-7793af0ad15e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.866271] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 635.866271] env[65788]: value = "task-4661804" [ 635.866271] env[65788]: _type = "Task" [ 635.866271] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.877078] env[65788]: INFO nova.compute.manager [-] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Took 1.89 seconds to deallocate network for instance. [ 635.877078] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661804, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.877770] env[65788]: DEBUG nova.network.neutron [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Successfully created port: 528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 635.887526] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 635.940053] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.940495] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.034944] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661802, 'name': PowerOffVM_Task, 'duration_secs': 0.229511} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.035348] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 636.035525] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 636.035784] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b3eef0a-fc7a-40b3-84b1-2263fe75cfa4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.067526] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 636.067860] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 636.068059] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleting the datastore file [datastore2] 39127362-8daf-43ff-99cb-ddb54c6fe095 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 636.068329] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14e9d429-083b-4afe-9b50-96a6ac31b4e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.075731] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for the task: (returnval){ [ 636.075731] env[65788]: value = "task-4661806" [ 636.075731] env[65788]: _type = "Task" [ 636.075731] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.084640] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.089474] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.089704] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.118199] env[65788]: INFO nova.compute.manager [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Took 15.69 seconds to build instance. [ 636.194112] env[65788]: DEBUG nova.scheduler.client.report [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 636.273314] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661803, 'name': ReconfigVM_Task, 'duration_secs': 0.453992} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.274309] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 636.379124] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661804, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.386777] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.416199] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.486799] env[65788]: DEBUG nova.network.neutron [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 636.547463] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.547463] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.587251] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.591613] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 636.622713] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3ea0f92-729c-439b-9c1b-28c0e3d4f80b tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "54b0945e-c9f5-4726-928e-de1d0ded5aae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.212s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.700578] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.006s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.703244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.209s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.704980] env[65788]: INFO nova.compute.claims [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.741970] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 636.751496] env[65788]: INFO nova.scheduler.client.report [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Deleted allocations for instance c06991cd-15b0-4d48-849a-140720ee07a0 [ 636.772934] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 636.773210] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 636.773347] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 636.773525] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 636.773663] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 636.773801] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 636.773998] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.774349] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 636.774709] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 636.774813] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 636.774981] env[65788]: DEBUG nova.virt.hardware [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 636.776098] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb4f3f6-8fb7-4d60-ac80-921f08950c5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.781098] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 636.781338] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 636.781482] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 636.781659] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 636.781797] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 636.781939] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 636.782291] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.782580] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 636.783264] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 636.783439] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 636.783991] env[65788]: DEBUG nova.virt.hardware [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 636.791034] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Reconfiguring VM instance instance-00000002 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 636.791034] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a5ba569-ccb9-4ec6-94fc-254803cdd3c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.813849] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec706fff-bfbd-4bc7-bc34-739a1701bbf6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.819061] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 636.819061] env[65788]: value = "task-4661807" [ 636.819061] env[65788]: _type = "Task" [ 636.819061] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.840479] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.877938] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661804, 'name': ReconfigVM_Task, 'duration_secs': 1.001683} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.878667] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 77dcec96-50e5-4753-b3cb-c6aec377bce2/77dcec96-50e5-4753-b3cb-c6aec377bce2.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 636.879660] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ae57c08-d9b3-45c7-b3e1-9c903afe0983 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.888137] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 636.888137] env[65788]: value = "task-4661808" [ 636.888137] env[65788]: _type = "Task" [ 636.888137] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.900341] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661808, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.921467] env[65788]: WARNING neutronclient.v2_0.client [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 636.921467] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.921467] env[65788]: WARNING openstack [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.090047] env[65788]: DEBUG oslo_vmware.api [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Task: {'id': task-4661806, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.609315} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.090346] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 637.090519] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 637.090684] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.090959] env[65788]: INFO nova.compute.manager [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Took 1.60 seconds to destroy the instance on the hypervisor. [ 637.091228] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 637.091422] env[65788]: DEBUG nova.compute.manager [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 637.091511] env[65788]: DEBUG nova.network.neutron [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 637.091754] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.092388] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.093114] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.108702] env[65788]: DEBUG nova.network.neutron [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Updating instance_info_cache with network_info: [{"id": "1b0ec4f3-be92-493e-a8ed-6af0f0bb844c", "address": "fa:16:3e:99:9f:a8", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0ec4f3-be", "ovs_interfaceid": "1b0ec4f3-be92-493e-a8ed-6af0f0bb844c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 637.132460] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.160753] env[65788]: DEBUG nova.network.neutron [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 637.161107] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 637.330257] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661807, 'name': ReconfigVM_Task, 'duration_secs': 0.20264} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.330497] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Reconfigured VM instance instance-00000002 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 637.331379] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4fed4d-ccbe-4db6-b284-9d91627ac3f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.360997] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] ed55713e-8bf1-4960-8309-47498b163d02/ed55713e-8bf1-4960-8309-47498b163d02.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 637.365272] env[65788]: DEBUG oslo_concurrency.lockutils [None req-05e8a4e2-ee8f-426d-b87a-0055f5be23b1 tempest-ImagesNegativeTestJSON-2006960124 tempest-ImagesNegativeTestJSON-2006960124-project-member] Lock "c06991cd-15b0-4d48-849a-140720ee07a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.965s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 637.366276] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c58e92d8-405d-40c7-bb79-577eea9aa575 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.389454] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 637.389454] env[65788]: value = "task-4661809" [ 637.389454] env[65788]: _type = "Task" [ 637.389454] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.403686] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661808, 'name': Rename_Task, 'duration_secs': 0.247848} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.412077] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 637.412485] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661809, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.413278] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c643563d-f598-4107-8153-58599db533fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.423767] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 637.423767] env[65788]: value = "task-4661810" [ 637.423767] env[65788]: _type = "Task" [ 637.423767] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.440405] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.569513] env[65788]: DEBUG nova.network.neutron [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Successfully updated port: 528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 637.616315] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "refresh_cache-5cef20ed-d9f0-4237-ae9d-da401b953904" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.616845] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance network_info: |[{"id": "1b0ec4f3-be92-493e-a8ed-6af0f0bb844c", "address": "fa:16:3e:99:9f:a8", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0ec4f3-be", "ovs_interfaceid": "1b0ec4f3-be92-493e-a8ed-6af0f0bb844c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 637.617671] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:9f:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b0ec4f3-be92-493e-a8ed-6af0f0bb844c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 637.627818] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating folder: Project (06b3dfec7c0144928b4a59dbd27569b7). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.631777] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-375ceea8-e13a-4c37-914e-cd8a118e2369 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.636683] env[65788]: INFO nova.compute.manager [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Rebuilding instance [ 637.648149] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created folder: Project (06b3dfec7c0144928b4a59dbd27569b7) in parent group-v910111. [ 637.648436] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating folder: Instances. Parent ref: group-v910157. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.653271] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e31fdee-0ac2-416b-9692-2622aaa8f900 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.665674] env[65788]: DEBUG nova.network.neutron [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 637.675210] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created folder: Instances in parent group-v910157. [ 637.675521] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 637.683338] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 637.683338] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c619f7f2-2fb5-4b7a-ae12-137722ab2239 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.711719] env[65788]: DEBUG nova.compute.manager [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 637.712908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997d5b40-3cd4-42dc-88fb-9da9a46668e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.724037] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 637.724037] env[65788]: value = "task-4661813" [ 637.724037] env[65788]: _type = "Task" [ 637.724037] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.742141] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661813, 'name': CreateVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.834787] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93487894-1552-43fd-9dd1-9a2127acb629 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.844586] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ffc08b-3360-4bb8-ab47-937d0f93a992 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.881047] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5b0cf0-a975-48c2-8b9d-c9149bc837d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.886173] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Received event network-changed-ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 637.886374] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Refreshing instance network info cache due to event network-changed-ca17fd35-e868-4f50-ac40-b6da3a12b015. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 637.886584] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Acquiring lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.887573] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Acquired lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.887573] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Refreshing network info cache for port ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 637.899202] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e63f32d-8c62-49b5-9c50-fbee36d5a6a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.924045] env[65788]: DEBUG nova.compute.provider_tree [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.924332] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661809, 'name': ReconfigVM_Task, 'duration_secs': 0.320357} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.924680] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Reconfigured VM instance instance-00000002 to attach disk [datastore2] ed55713e-8bf1-4960-8309-47498b163d02/ed55713e-8bf1-4960-8309-47498b163d02.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 637.924927] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 637.940285] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661810, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.033873] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.034251] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.077893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "refresh_cache-15aba949-c04c-4021-add8-2e9d4e4771db" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.078304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquired lock "refresh_cache-15aba949-c04c-4021-add8-2e9d4e4771db" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.078550] env[65788]: DEBUG nova.network.neutron [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 638.175116] env[65788]: INFO nova.compute.manager [-] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Took 1.08 seconds to deallocate network for instance. [ 638.235293] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661813, 'name': CreateVM_Task, 'duration_secs': 0.435604} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.237305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 638.238089] env[65788]: WARNING neutronclient.v2_0.client [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.238455] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.238616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.239045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 638.240405] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2df006f6-5a76-41b3-b9a3-e188aedb210d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.246087] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 638.246087] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c65436-6e21-17d8-40e9-b556d2d56cf4" [ 638.246087] env[65788]: _type = "Task" [ 638.246087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.257202] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c65436-6e21-17d8-40e9-b556d2d56cf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.329194] env[65788]: DEBUG nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Received event network-vif-plugged-1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 638.329411] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Acquiring lock "5cef20ed-d9f0-4237-ae9d-da401b953904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.329411] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.329581] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.329738] env[65788]: DEBUG nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] No waiting events found dispatching network-vif-plugged-1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 638.329943] env[65788]: WARNING nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Received unexpected event network-vif-plugged-1b0ec4f3-be92-493e-a8ed-6af0f0bb844c for instance with vm_state building and task_state spawning. [ 638.330127] env[65788]: DEBUG nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Received event network-changed-1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 638.330276] env[65788]: DEBUG nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Refreshing instance network info cache due to event network-changed-1b0ec4f3-be92-493e-a8ed-6af0f0bb844c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 638.330452] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Acquiring lock "refresh_cache-5cef20ed-d9f0-4237-ae9d-da401b953904" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.330581] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Acquired lock "refresh_cache-5cef20ed-d9f0-4237-ae9d-da401b953904" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.331203] env[65788]: DEBUG nova.network.neutron [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Refreshing network info cache for port 1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 638.392775] env[65788]: WARNING neutronclient.v2_0.client [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.393554] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.393918] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.430382] env[65788]: DEBUG nova.scheduler.client.report [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 638.436537] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878c0a6b-489d-4503-9371-94028a6044f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.442630] env[65788]: DEBUG oslo_vmware.api [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4661810, 'name': PowerOnVM_Task, 'duration_secs': 0.650321} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.443265] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 638.443472] env[65788]: INFO nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Took 11.81 seconds to spawn the instance on the hypervisor. [ 638.443863] env[65788]: DEBUG nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 638.444742] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b02b699-d1ea-4871-a6c9-6365148deff2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.462965] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083426c2-83c1-4090-a186-f33a7a3fb93e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.488428] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 638.537246] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 638.581589] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.581985] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.681992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.740813] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 638.741383] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca2c1912-3104-42d5-b173-dc273b3e8ed4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.753485] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 638.753485] env[65788]: value = "task-4661814" [ 638.753485] env[65788]: _type = "Task" [ 638.753485] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.763185] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c65436-6e21-17d8-40e9-b556d2d56cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.017434} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.763185] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.763185] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 638.763185] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.763467] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.763467] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 638.763467] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9467f12d-04ca-42e2-a8c0-67634d97732b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.771292] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.781161] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 638.781350] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 638.783021] env[65788]: DEBUG nova.network.neutron [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 638.784876] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2737ae6e-f8f8-4c7c-adb8-c77a3d11eb40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.792671] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 638.792671] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52df9552-d1e6-1e0b-e65b-b5b7abbb549f" [ 638.792671] env[65788]: _type = "Task" [ 638.792671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.803540] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52df9552-d1e6-1e0b-e65b-b5b7abbb549f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.831633] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.834282] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.849115] env[65788]: WARNING neutronclient.v2_0.client [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.853127] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.853127] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.861105] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.861359] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.937346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.234s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.938384] env[65788]: DEBUG nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 638.941821] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.106s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.942375] env[65788]: DEBUG nova.objects.instance [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lazy-loading 'resources' on Instance uuid 98a2efec-cb20-4e95-9ed5-5b2057a48c39 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 638.953733] env[65788]: WARNING neutronclient.v2_0.client [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.954725] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.955286] env[65788]: WARNING openstack [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.969615] env[65788]: WARNING neutronclient.v2_0.client [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 638.973077] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.973077] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.989608] env[65788]: INFO nova.compute.manager [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Took 19.31 seconds to build instance. [ 638.994807] env[65788]: WARNING neutronclient.v2_0.client [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.069662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.116488] env[65788]: DEBUG nova.network.neutron [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Port ffa53d50-ad0e-438c-994a-b0b7a97f85f2 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 639.137309] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.137309] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.179150] env[65788]: DEBUG nova.network.neutron [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Updating instance_info_cache with network_info: [{"id": "528d3b24-3631-4afa-af44-515f60a2df2c", "address": "fa:16:3e:94:97:5e", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.44", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap528d3b24-36", "ovs_interfaceid": "528d3b24-3631-4afa-af44-515f60a2df2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.197154] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Updated VIF entry in instance network info cache for port ca17fd35-e868-4f50-ac40-b6da3a12b015. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 639.197978] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Updating instance_info_cache with network_info: [{"id": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "address": "fa:16:3e:7e:b8:93", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca17fd35-e8", "ovs_interfaceid": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.228832] env[65788]: WARNING neutronclient.v2_0.client [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 639.229011] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.231435] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.270901] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661814, 'name': PowerOffVM_Task, 'duration_secs': 0.329503} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.270901] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 639.270901] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.270901] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130f6712-ed25-4301-afe3-0498fb55d0c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.279532] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 639.279984] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a300602-1967-4cf3-90e2-445a414de933 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.306221] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52df9552-d1e6-1e0b-e65b-b5b7abbb549f, 'name': SearchDatastore_Task, 'duration_secs': 0.018355} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.307271] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-030ef92d-ac4f-40ae-8e8a-9ad36169d18c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.315058] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 639.315058] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524f1c39-fbf1-131f-f1da-2661dcba1993" [ 639.315058] env[65788]: _type = "Task" [ 639.315058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.327311] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524f1c39-fbf1-131f-f1da-2661dcba1993, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.338433] env[65788]: DEBUG nova.network.neutron [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Updated VIF entry in instance network info cache for port 1b0ec4f3-be92-493e-a8ed-6af0f0bb844c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 639.338804] env[65788]: DEBUG nova.network.neutron [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Updating instance_info_cache with network_info: [{"id": "1b0ec4f3-be92-493e-a8ed-6af0f0bb844c", "address": "fa:16:3e:99:9f:a8", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b0ec4f3-be", "ovs_interfaceid": "1b0ec4f3-be92-493e-a8ed-6af0f0bb844c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 639.364750] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 639.365024] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 639.365315] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore1] 141aefe8-1b95-4963-854d-da79ddf143f7 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 639.365598] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b3e3935-2177-452e-8cec-62f6fd6b0d57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.374668] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 639.374668] env[65788]: value = "task-4661816" [ 639.374668] env[65788]: _type = "Task" [ 639.374668] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.386945] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.450203] env[65788]: DEBUG nova.compute.utils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 639.451373] env[65788]: DEBUG nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 639.494663] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4628dffb-d6cc-4fd6-b38d-148f07946953 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.836s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.682371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Releasing lock "refresh_cache-15aba949-c04c-4021-add8-2e9d4e4771db" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.683565] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Instance network_info: |[{"id": "528d3b24-3631-4afa-af44-515f60a2df2c", "address": "fa:16:3e:94:97:5e", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.44", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap528d3b24-36", "ovs_interfaceid": "528d3b24-3631-4afa-af44-515f60a2df2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 639.683877] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:97:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '528d3b24-3631-4afa-af44-515f60a2df2c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.692498] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Creating folder: Project (3dc8fe6cf1ed4cdb91bf077edabce4b8). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.696204] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f6a49f9-fe14-440f-bdc2-b37609b04580 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.702112] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Releasing lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.702387] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Received event network-vif-deleted-5cc1dfe1-7bda-4f6a-9828-254a5f93a508 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 639.702557] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Received event network-changed-709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 639.702717] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Refreshing instance network info cache due to event network-changed-709378e7-1b93-4d0c-9649-3fb90a38b091. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 639.702906] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Acquiring lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.703092] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Acquired lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.703199] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Refreshing network info cache for port 709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 639.709232] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Created folder: Project (3dc8fe6cf1ed4cdb91bf077edabce4b8) in parent group-v910111. [ 639.709232] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Creating folder: Instances. Parent ref: group-v910160. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.712907] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67f389cc-8193-4b8b-947d-12b7c2ca800a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.727568] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Created folder: Instances in parent group-v910160. [ 639.727917] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 639.728199] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 639.728805] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-716ad619-b459-481c-8f69-4dfc67ece01e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.755203] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.755203] env[65788]: value = "task-4661819" [ 639.755203] env[65788]: _type = "Task" [ 639.755203] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.765778] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661819, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.829175] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524f1c39-fbf1-131f-f1da-2661dcba1993, 'name': SearchDatastore_Task, 'duration_secs': 0.027036} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.829467] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.829719] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 639.830034] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e4495db-1590-49fa-8ce5-32ff8de0dae5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.834701] env[65788]: INFO nova.compute.manager [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Rebuilding instance [ 639.838423] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 639.838423] env[65788]: value = "task-4661820" [ 639.838423] env[65788]: _type = "Task" [ 639.838423] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.842580] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Releasing lock "refresh_cache-5cef20ed-d9f0-4237-ae9d-da401b953904" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.842709] env[65788]: DEBUG nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Received event network-changed-d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 639.842884] env[65788]: DEBUG nova.compute.manager [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Refreshing instance network info cache due to event network-changed-d6e04a22-9f86-43c9-ac42-95191eaf454b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 639.843109] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.843248] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.843419] env[65788]: DEBUG nova.network.neutron [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Refreshing network info cache for port d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 639.853991] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661820, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.886692] env[65788]: DEBUG nova.compute.manager [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 639.887591] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e77671-4839-4e67-9e6f-8803e1f24b70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.897028] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352213} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.900548] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 639.900883] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 639.901144] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.919225] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5fd2e2-a90f-41eb-8230-c41f751c9090 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.927861] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eea409e-db13-48fb-96c5-cf54e79edead {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.964303] env[65788]: DEBUG nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 639.971039] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f191838-6084-4c39-9338-c837a116a97e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.978499] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68809ce2-488f-446d-9bb1-1ca39875c189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.994591] env[65788]: DEBUG nova.compute.provider_tree [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.145730] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "ed55713e-8bf1-4960-8309-47498b163d02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.145730] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.145926] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.206879] env[65788]: WARNING neutronclient.v2_0.client [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.207164] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.207545] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.279841] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661819, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.352813] env[65788]: WARNING neutronclient.v2_0.client [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.353612] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.354043] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.361904] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661820, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.372569] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.373040] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.498954] env[65788]: DEBUG nova.scheduler.client.report [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 640.587820] env[65788]: WARNING neutronclient.v2_0.client [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 640.588525] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.588879] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.663022] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.663022] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.770040] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661819, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.851076] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661820, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.770529} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.851372] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 640.851577] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 640.851858] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb48e78d-c353-44f2-b317-2a5dafc4284d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.859912] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 640.859912] env[65788]: value = "task-4661821" [ 640.859912] env[65788]: _type = "Task" [ 640.859912] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.869632] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661821, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.918550] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 640.920132] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79739b46-a99b-402f-acad-c91337531946 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.929525] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 640.929525] env[65788]: value = "task-4661822" [ 640.929525] env[65788]: _type = "Task" [ 640.929525] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.942848] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.945849] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 640.946098] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 640.946254] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 640.946470] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 640.946624] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 640.946692] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 640.946884] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.947175] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 640.947396] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 640.947617] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 640.947917] env[65788]: DEBUG nova.virt.hardware [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 640.948718] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927577ed-489e-4287-b940-0563a30007bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.958694] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5315bc99-6d31-4cc6-b056-4c0f01070fb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.977807] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:71:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd564449f-c908-4d2b-9416-3ea9d3d371fc', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 640.987471] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 640.988704] env[65788]: DEBUG nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 640.991901] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 640.992418] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-604e4763-58cc-469e-a1ca-2c5ac228834a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.014948] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updated VIF entry in instance network info cache for port 709378e7-1b93-4d0c-9649-3fb90a38b091. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 641.015407] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updating instance_info_cache with network_info: [{"id": "709378e7-1b93-4d0c-9649-3fb90a38b091", "address": "fa:16:3e:8a:05:2d", "network": {"id": "a8e54c54-a388-46c7-b7bd-556b975e8abd", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-501645486-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.235", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4fa459531885434f9a977d68864b7042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap709378e7-1b", "ovs_interfaceid": "709378e7-1b93-4d0c-9649-3fb90a38b091", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.017186] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.023159] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.692s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.030056] env[65788]: INFO nova.compute.claims [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.040251] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 641.040251] env[65788]: value = "task-4661823" [ 641.040251] env[65788]: _type = "Task" [ 641.040251] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.044125] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 641.044704] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 641.045049] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 641.045371] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 641.046272] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 641.046272] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 641.046593] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.046912] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 641.047973] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 641.047973] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 641.047973] env[65788]: DEBUG nova.virt.hardware [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 641.048844] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef575cf0-1683-4ba7-9079-1ed903fd6f89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.053711] env[65788]: WARNING neutronclient.v2_0.client [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.054447] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.055776] env[65788]: WARNING openstack [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.071374] env[65788]: INFO nova.scheduler.client.report [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Deleted allocations for instance 98a2efec-cb20-4e95-9ed5-5b2057a48c39 [ 641.084254] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661823, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.087417] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b6c800-0e59-4746-8ccf-f13bf4514e43 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.110136] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 641.118563] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 641.120225] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 641.120225] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5579e62-44fe-44b1-8b7b-4aaf435505ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.148081] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 641.148081] env[65788]: value = "task-4661824" [ 641.148081] env[65788]: _type = "Task" [ 641.148081] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.157078] env[65788]: WARNING neutronclient.v2_0.client [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.158221] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661824, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.270260] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.270511] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.270697] env[65788]: DEBUG nova.network.neutron [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 641.281939] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661819, 'name': CreateVM_Task, 'duration_secs': 1.387064} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.288126] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 641.288853] env[65788]: WARNING neutronclient.v2_0.client [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.289781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.289781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.289928] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 641.290219] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58a38695-b772-4bf8-832b-f14a49cf57f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.297749] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 641.297749] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204e00e-068f-b362-2405-b6e16a2ae80b" [ 641.297749] env[65788]: _type = "Task" [ 641.297749] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.309971] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204e00e-068f-b362-2405-b6e16a2ae80b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.318330] env[65788]: DEBUG nova.network.neutron [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updated VIF entry in instance network info cache for port d6e04a22-9f86-43c9-ac42-95191eaf454b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 641.319037] env[65788]: DEBUG nova.network.neutron [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 641.376179] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661821, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072575} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.376502] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 641.377699] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6636c267-ab78-459e-abed-0e76ce5e2ae9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.405507] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 641.405883] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bd2a2cd-299d-4523-b129-07beb19967af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.428812] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 641.428812] env[65788]: value = "task-4661825" [ 641.428812] env[65788]: _type = "Task" [ 641.428812] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.446193] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661822, 'name': PowerOffVM_Task, 'duration_secs': 0.183666} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.446470] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.446706] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 641.447052] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.447769] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090b0a07-650f-47fb-a98a-44e513d7a226 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.457493] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 641.457921] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b63b19e-576e-4b47-936e-a212b95860fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.484956] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 641.485370] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 641.485370] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Deleting the datastore file [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 641.486036] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2286b5e-739c-47b5-9a01-40eef2593d57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.493536] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 641.493536] env[65788]: value = "task-4661827" [ 641.493536] env[65788]: _type = "Task" [ 641.493536] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.501755] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.522626] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Releasing lock "refresh_cache-5b701040-025c-4246-ad54-f2cf478e998d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.522976] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Received event network-changed-e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 641.523160] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Refreshing instance network info cache due to event network-changed-e936d0e7-de51-4ac2-9b4a-08f999624c5c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 641.523300] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Acquiring lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.523472] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Acquired lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.523643] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Refreshing network info cache for port e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 641.558748] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661823, 'name': CreateVM_Task, 'duration_secs': 0.407406} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.558923] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 641.559651] env[65788]: WARNING neutronclient.v2_0.client [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 641.559788] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.588518] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8dd0c206-ccc5-4011-a1eb-da64b2decb3b tempest-ServerDiagnosticsTest-1881811713 tempest-ServerDiagnosticsTest-1881811713-project-member] Lock "98a2efec-cb20-4e95-9ed5-5b2057a48c39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.770s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.658279] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661824, 'name': CreateVM_Task, 'duration_secs': 0.321391} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.658279] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 641.658279] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.724697] env[65788]: DEBUG nova.compute.manager [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Received event network-vif-plugged-528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 641.724697] env[65788]: DEBUG oslo_concurrency.lockutils [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Acquiring lock "15aba949-c04c-4021-add8-2e9d4e4771db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.725357] env[65788]: DEBUG oslo_concurrency.lockutils [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Lock "15aba949-c04c-4021-add8-2e9d4e4771db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.725799] env[65788]: DEBUG oslo_concurrency.lockutils [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Lock "15aba949-c04c-4021-add8-2e9d4e4771db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.726415] env[65788]: DEBUG nova.compute.manager [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] No waiting events found dispatching network-vif-plugged-528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 641.726415] env[65788]: WARNING nova.compute.manager [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Received unexpected event network-vif-plugged-528d3b24-3631-4afa-af44-515f60a2df2c for instance with vm_state building and task_state spawning. [ 641.726415] env[65788]: DEBUG nova.compute.manager [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Received event network-changed-528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 641.726760] env[65788]: DEBUG nova.compute.manager [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Refreshing instance network info cache due to event network-changed-528d3b24-3631-4afa-af44-515f60a2df2c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 641.727036] env[65788]: DEBUG oslo_concurrency.lockutils [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Acquiring lock "refresh_cache-15aba949-c04c-4021-add8-2e9d4e4771db" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.727124] env[65788]: DEBUG oslo_concurrency.lockutils [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Acquired lock "refresh_cache-15aba949-c04c-4021-add8-2e9d4e4771db" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.727581] env[65788]: DEBUG nova.network.neutron [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Refreshing network info cache for port 528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 641.775343] env[65788]: WARNING neutronclient.v2_0.client [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 641.776149] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.776546] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.811814] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204e00e-068f-b362-2405-b6e16a2ae80b, 'name': SearchDatastore_Task, 'duration_secs': 0.020815} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.812137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.813381] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 641.813517] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.813702] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.813893] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.815294] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.816699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 641.816699] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8f37b4b-b554-455e-8611-e298709c0be6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.817865] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b2c9a50-1ae1-476b-9154-63a247a89114 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.821187] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea2fede9-ec9a-49a7-9a43-70f3b0b8540b req-f83d7c57-0fbe-4362-91a4-bf92e8191cd4 service nova] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.825268] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 641.825268] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5222402d-4f28-5c90-48de-01ce62754d42" [ 641.825268] env[65788]: _type = "Task" [ 641.825268] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.829850] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.830110] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 641.833699] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f010b96-2fb9-44db-8aea-195cd43b5e1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.841166] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5222402d-4f28-5c90-48de-01ce62754d42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.844802] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 641.844802] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b97b5-cb4c-4aa9-4d2d-fbfa42878804" [ 641.844802] env[65788]: _type = "Task" [ 641.844802] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.853456] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b97b5-cb4c-4aa9-4d2d-fbfa42878804, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.939465] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.007278] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298071} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.007527] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 642.007699] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 642.007883] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 642.026908] env[65788]: WARNING neutronclient.v2_0.client [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.028943] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.028943] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.231687] env[65788]: WARNING neutronclient.v2_0.client [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.232475] env[65788]: WARNING openstack [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.233242] env[65788]: WARNING openstack [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.344468] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5222402d-4f28-5c90-48de-01ce62754d42, 'name': SearchDatastore_Task, 'duration_secs': 0.015568} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.344925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.345297] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 642.345726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.345843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.346143] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 642.352599] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab799682-4c6a-455d-851b-2bf6d309fa5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.368579] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b97b5-cb4c-4aa9-4d2d-fbfa42878804, 'name': SearchDatastore_Task, 'duration_secs': 0.021305} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.369256] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 642.369256] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521548c3-ebed-bff0-5b64-95c4cd52b630" [ 642.369256] env[65788]: _type = "Task" [ 642.369256] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.370109] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b00c0292-65bc-4e54-83c5-239d9b22b51a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.383301] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 642.383301] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c43428-26c2-b14f-102c-079e4503442a" [ 642.383301] env[65788]: _type = "Task" [ 642.383301] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.389538] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521548c3-ebed-bff0-5b64-95c4cd52b630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.402691] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c43428-26c2-b14f-102c-079e4503442a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.418628] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.421022] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.443370] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661825, 'name': ReconfigVM_Task, 'duration_secs': 0.812783} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.443671] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Reconfigured VM instance instance-00000010 to attach disk [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 642.444435] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87d0a924-a0c3-46aa-ab1b-bd66fd4560f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.454676] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 642.454676] env[65788]: value = "task-4661828" [ 642.454676] env[65788]: _type = "Task" [ 642.454676] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.461408] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd56779-3ef6-4248-aa11-50c363ea0876 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.467584] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661828, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.473548] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdcd772-41d0-4481-b236-9c9d813cf21e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.521781] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d33fbbb-8418-42ce-a3c5-7b76033c014c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.531336] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c597c9-1c27-445d-ad84-9d38a8e1f4e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.547695] env[65788]: DEBUG nova.compute.provider_tree [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 642.623946] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.624285] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.798116] env[65788]: WARNING neutronclient.v2_0.client [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 642.800207] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.800207] env[65788]: WARNING openstack [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.810592] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.811070] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.818312] env[65788]: WARNING openstack [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.818677] env[65788]: WARNING openstack [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.885450] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521548c3-ebed-bff0-5b64-95c4cd52b630, 'name': SearchDatastore_Task, 'duration_secs': 0.036385} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.885450] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.885621] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 642.885830] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.902230] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c43428-26c2-b14f-102c-079e4503442a, 'name': SearchDatastore_Task, 'duration_secs': 0.027411} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.905018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.905018] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 15aba949-c04c-4021-add8-2e9d4e4771db/15aba949-c04c-4021-add8-2e9d4e4771db.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 642.905018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.905018] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 642.905322] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d418081-21b4-4327-ab5d-bc529ed7d055 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.905627] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2d1e6a8-a2d8-48b6-938f-fc37295f4982 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.914609] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 642.914609] env[65788]: value = "task-4661829" [ 642.914609] env[65788]: _type = "Task" [ 642.914609] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.923169] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 642.923353] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 642.924729] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f463c6fe-67d2-4d63-969c-afbf28bbb6ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.930827] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661829, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.935441] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 642.935441] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528c986f-d67e-e053-318f-4ce7484e246b" [ 642.935441] env[65788]: _type = "Task" [ 642.935441] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.945173] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528c986f-d67e-e053-318f-4ce7484e246b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.965626] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661828, 'name': Rename_Task, 'duration_secs': 0.288042} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.966338] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 642.966338] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13252f4f-68d2-49d6-b4d1-3be19bab691b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.973808] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 642.973808] env[65788]: value = "task-4661830" [ 642.973808] env[65788]: _type = "Task" [ 642.973808] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.984561] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661830, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.022950] env[65788]: DEBUG nova.network.neutron [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [{"id": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "address": "fa:16:3e:2c:93:04", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa53d50-ad", "ovs_interfaceid": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 643.060717] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 643.063897] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 643.063897] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 643.063897] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 643.063897] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 643.063897] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 643.063897] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.064238] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 643.064238] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 643.064238] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 643.064238] env[65788]: DEBUG nova.virt.hardware [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 643.064551] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce52a4c1-e440-414a-af06-04b43773a5f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.074617] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9828a2-901f-4589-b565-7e131cbff391 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.082338] env[65788]: WARNING neutronclient.v2_0.client [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.083057] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.083404] env[65788]: WARNING openstack [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.102084] env[65788]: ERROR nova.scheduler.client.report [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [req-7556d64d-e018-477f-8fcf-b635e150d0b4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7556d64d-e018-477f-8fcf-b635e150d0b4"}]} [ 643.102935] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 643.108803] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 643.112317] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 643.112317] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a4381d8-0422-447a-9073-508016041736 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.126498] env[65788]: WARNING neutronclient.v2_0.client [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.127493] env[65788]: WARNING openstack [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.128119] env[65788]: WARNING openstack [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.137381] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 643.140771] env[65788]: DEBUG nova.scheduler.client.report [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 643.149595] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 643.149595] env[65788]: value = "task-4661831" [ 643.149595] env[65788]: _type = "Task" [ 643.149595] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.159884] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661831, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.170213] env[65788]: DEBUG nova.scheduler.client.report [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 643.171047] env[65788]: DEBUG nova.compute.provider_tree [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 643.189296] env[65788]: DEBUG nova.scheduler.client.report [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 643.218985] env[65788]: DEBUG nova.scheduler.client.report [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 643.306179] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updated VIF entry in instance network info cache for port e936d0e7-de51-4ac2-9b4a-08f999624c5c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 643.306179] env[65788]: DEBUG nova.network.neutron [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updating instance_info_cache with network_info: [{"id": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "address": "fa:16:3e:ff:eb:0c", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape936d0e7-de", "ovs_interfaceid": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 643.345218] env[65788]: DEBUG nova.network.neutron [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Updated VIF entry in instance network info cache for port 528d3b24-3631-4afa-af44-515f60a2df2c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 643.346281] env[65788]: DEBUG nova.network.neutron [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Updating instance_info_cache with network_info: [{"id": "528d3b24-3631-4afa-af44-515f60a2df2c", "address": "fa:16:3e:94:97:5e", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.44", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap528d3b24-36", "ovs_interfaceid": "528d3b24-3631-4afa-af44-515f60a2df2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 643.433248] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661829, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.450505] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528c986f-d67e-e053-318f-4ce7484e246b, 'name': SearchDatastore_Task, 'duration_secs': 0.011943} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.451405] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d07120ba-7ed4-42fa-9395-ac17ace4f880 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.458364] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 643.458364] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed2aa9-8e99-c7f3-b293-5b2d192314fe" [ 643.458364] env[65788]: _type = "Task" [ 643.458364] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.472278] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed2aa9-8e99-c7f3-b293-5b2d192314fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.488136] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661830, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.530132] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.669445] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661831, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.671505] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.706015] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4600005b-a10d-489e-9e50-37b33696fb86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.714243] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ab397b-1403-450b-8146-1253a87c6aab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.748572] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec5ae60-2bee-41bc-843d-4a26aaf9b595 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.757191] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de8ae11-2e6c-42fd-b486-0deeee6ad534 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.773958] env[65788]: DEBUG nova.compute.provider_tree [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.811864] env[65788]: DEBUG oslo_concurrency.lockutils [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] Releasing lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.814227] env[65788]: DEBUG nova.compute.manager [req-002cf1fa-3db2-40b7-be54-5592c554d443 req-7edd4f66-97dd-4eba-ae25-e4fd6e7c5cbe service nova] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Received event network-vif-deleted-a0bbf2f7-7442-4457-a3e7-eba29d08da68 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 643.849111] env[65788]: DEBUG oslo_concurrency.lockutils [req-aa8cf454-9e2a-415b-a5db-543b974bfd7a req-7c9ab684-474f-471c-b605-566204b11af1 service nova] Releasing lock "refresh_cache-15aba949-c04c-4021-add8-2e9d4e4771db" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.886319] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "350f3c86-063c-4ed5-895b-fc621b2fa825" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.888019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.888019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "350f3c86-063c-4ed5-895b-fc621b2fa825-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.888019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.888019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.889838] env[65788]: INFO nova.compute.manager [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Terminating instance [ 643.930011] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661829, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.600834} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.930315] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 15aba949-c04c-4021-add8-2e9d4e4771db/15aba949-c04c-4021-add8-2e9d4e4771db.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 643.930606] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 643.930772] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29d51c24-7b7c-400b-aaa6-54e9ccfd1339 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.941058] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 643.941058] env[65788]: value = "task-4661832" [ 643.941058] env[65788]: _type = "Task" [ 643.941058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.954222] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.971341] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed2aa9-8e99-c7f3-b293-5b2d192314fe, 'name': SearchDatastore_Task, 'duration_secs': 0.041625} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.971920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.971920] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 643.972117] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 643.972636] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 643.972636] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18dc1aa7-e068-42a5-9f41-27cc71475a37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.974860] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fb6d636-476f-4651-8031-c3c4b5b2c215 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.983877] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 643.983877] env[65788]: value = "task-4661833" [ 643.983877] env[65788]: _type = "Task" [ 643.983877] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.987194] env[65788]: DEBUG oslo_vmware.api [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661830, 'name': PowerOnVM_Task, 'duration_secs': 0.769587} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.991545] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 643.991545] env[65788]: INFO nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Took 10.34 seconds to spawn the instance on the hypervisor. [ 643.991545] env[65788]: DEBUG nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 643.991799] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 643.991946] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 643.993147] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0e6399-5bbb-4f2c-afb2-971868eaf31d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.995738] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a0b797b-565e-42fb-9264-e128d515fedc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.008173] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 644.008173] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a7cb88-8ce9-87a9-36fc-99889cf67691" [ 644.008173] env[65788]: _type = "Task" [ 644.008173] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.011699] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.021312] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a7cb88-8ce9-87a9-36fc-99889cf67691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.071201] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b0bf8e-9ee3-48cc-afbd-55b751f62df6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.098354] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b322ac-74d6-4c39-a9c4-3764e50124b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.107451] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 644.163595] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661831, 'name': CreateVM_Task, 'duration_secs': 0.598863} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.163778] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 644.164238] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.164346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.164691] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 644.164972] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5274780-d5bf-48fb-bf9f-71b754cb1d09 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.171088] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 644.171088] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5279f222-4819-9a08-35e8-b8c985a9c4d6" [ 644.171088] env[65788]: _type = "Task" [ 644.171088] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.183779] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5279f222-4819-9a08-35e8-b8c985a9c4d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.277702] env[65788]: DEBUG nova.scheduler.client.report [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 644.399422] env[65788]: DEBUG nova.compute.manager [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 644.399656] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.402032] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef4bf5f-9dce-4309-9da2-ba1603c582f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.409806] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 644.410147] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06a1ad5d-b889-4f46-b75f-b103fbdba321 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.419197] env[65788]: DEBUG oslo_vmware.api [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 644.419197] env[65788]: value = "task-4661834" [ 644.419197] env[65788]: _type = "Task" [ 644.419197] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.429379] env[65788]: DEBUG oslo_vmware.api [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661834, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.452473] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071702} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.452801] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 644.453532] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ad74b3-fbac-4d2a-a0b3-3602cbb8d0bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.477843] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] 15aba949-c04c-4021-add8-2e9d4e4771db/15aba949-c04c-4021-add8-2e9d4e4771db.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 644.478156] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58ae35e4-00de-4ac3-8385-7a2347c3349b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.506221] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.507741] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 644.507741] env[65788]: value = "task-4661835" [ 644.507741] env[65788]: _type = "Task" [ 644.507741] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.525363] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661835, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.526180] env[65788]: INFO nova.compute.manager [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Took 20.53 seconds to build instance. [ 644.533132] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a7cb88-8ce9-87a9-36fc-99889cf67691, 'name': SearchDatastore_Task, 'duration_secs': 0.032732} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.535101] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f84ba978-c869-4616-bcfd-36dfaed9a288 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.542692] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 644.542692] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52755091-3813-f8ad-73fb-d7a5129bda2b" [ 644.542692] env[65788]: _type = "Task" [ 644.542692] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.555171] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52755091-3813-f8ad-73fb-d7a5129bda2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.617738] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.617738] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ca2b791-6b89-42a7-a0b2-8bb4e4cf71d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.627304] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 644.627304] env[65788]: value = "task-4661836" [ 644.627304] env[65788]: _type = "Task" [ 644.627304] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.640100] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661836, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.686087] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5279f222-4819-9a08-35e8-b8c985a9c4d6, 'name': SearchDatastore_Task, 'duration_secs': 0.054399} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.686087] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 644.686087] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 644.686087] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.788366] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.766s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.788366] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 644.791667] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.418s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.794500] env[65788]: INFO nova.compute.claims [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 644.932734] env[65788]: DEBUG oslo_vmware.api [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661834, 'name': PowerOffVM_Task, 'duration_secs': 0.362209} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.933580] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 644.933580] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 644.933580] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3742b2f6-c922-4e5a-8961-ffe2e9b6574b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.009115] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661833, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.949564} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.014414] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 645.014686] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 645.015793] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba65cf23-5334-4527-b66a-39044a499a56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.019526] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 645.019779] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 645.019988] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Deleting the datastore file [datastore2] 350f3c86-063c-4ed5-895b-fc621b2fa825 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 645.023920] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a4d686f-1bef-4d97-a2aa-8df6eb98a5cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.026248] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661835, 'name': ReconfigVM_Task, 'duration_secs': 0.474413} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.027762] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Reconfigured VM instance instance-00000011 to attach disk [datastore2] 15aba949-c04c-4021-add8-2e9d4e4771db/15aba949-c04c-4021-add8-2e9d4e4771db.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 645.028545] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 645.028545] env[65788]: value = "task-4661838" [ 645.028545] env[65788]: _type = "Task" [ 645.028545] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.029141] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3b9776b-b7a7-484a-a099-2fd344e5f5c1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.038028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-853033fb-71d5-4354-b4a4-1cfabfcd77d5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.051s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.038806] env[65788]: DEBUG oslo_vmware.api [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for the task: (returnval){ [ 645.038806] env[65788]: value = "task-4661839" [ 645.038806] env[65788]: _type = "Task" [ 645.038806] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.047680] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661838, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.048121] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 645.048121] env[65788]: value = "task-4661840" [ 645.048121] env[65788]: _type = "Task" [ 645.048121] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.059615] env[65788]: DEBUG oslo_vmware.api [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661839, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.067552] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52755091-3813-f8ad-73fb-d7a5129bda2b, 'name': SearchDatastore_Task, 'duration_secs': 0.014594} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.070829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.071260] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 645.071545] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661840, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.071796] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.071994] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 645.072303] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7df3bc42-d6d3-458f-bc05-355f15abf4cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.075885] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d96aff4b-18e6-4ef7-8fbe-2b5fa00e0443 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.084090] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 645.084090] env[65788]: value = "task-4661841" [ 645.084090] env[65788]: _type = "Task" [ 645.084090] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.085929] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 645.088190] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 645.092683] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-496fdc14-80e1-45fc-bdbf-8271fc856b77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.102184] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661841, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.104200] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 645.104200] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c4ec85-8f36-6614-35df-a633f551506c" [ 645.104200] env[65788]: _type = "Task" [ 645.104200] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.116344] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c4ec85-8f36-6614-35df-a633f551506c, 'name': SearchDatastore_Task, 'duration_secs': 0.012345} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.117474] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ae4c5a3-63e1-41b0-ae9c-0a90f4b4a45c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.127496] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 645.127496] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521833da-520e-4de5-1ed4-b82ce3634b47" [ 645.127496] env[65788]: _type = "Task" [ 645.127496] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.147914] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521833da-520e-4de5-1ed4-b82ce3634b47, 'name': SearchDatastore_Task, 'duration_secs': 0.014313} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.147914] env[65788]: DEBUG oslo_vmware.api [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4661836, 'name': PowerOnVM_Task, 'duration_secs': 0.492323} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.147914] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.147914] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 645.148098] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.148098] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2ee38883-8d2e-430a-b90e-e54c3b2eef9c tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance 'ed55713e-8bf1-4960-8309-47498b163d02' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 645.151759] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f2c5a3cf-8983-4c90-b847-d41f15c0c9ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.159883] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 645.159883] env[65788]: value = "task-4661842" [ 645.159883] env[65788]: _type = "Task" [ 645.159883] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.170054] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.208456] env[65788]: DEBUG nova.compute.manager [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Received event network-changed-ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 645.208678] env[65788]: DEBUG nova.compute.manager [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Refreshing instance network info cache due to event network-changed-ca17fd35-e868-4f50-ac40-b6da3a12b015. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 645.208893] env[65788]: DEBUG oslo_concurrency.lockutils [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Acquiring lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.213996] env[65788]: DEBUG oslo_concurrency.lockutils [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Acquired lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 645.213996] env[65788]: DEBUG nova.network.neutron [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Refreshing network info cache for port ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 645.299468] env[65788]: DEBUG nova.compute.utils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 645.303611] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 645.303856] env[65788]: DEBUG nova.network.neutron [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 645.306265] env[65788]: WARNING neutronclient.v2_0.client [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.306265] env[65788]: WARNING neutronclient.v2_0.client [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.306742] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.306742] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.360828] env[65788]: DEBUG nova.policy [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c028d9c053054a2f916507b237d37e60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06b3dfec7c0144928b4a59dbd27569b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 645.546944] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661838, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.547431] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 645.548593] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c97c0e-6e45-4f19-8f2e-988f25c7b278 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.562550] env[65788]: DEBUG oslo_vmware.api [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Task: {'id': task-4661839, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.36952} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.574796] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 645.575155] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 645.575405] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.575667] env[65788]: INFO nova.compute.manager [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Took 1.18 seconds to destroy the instance on the hypervisor. [ 645.575934] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 645.585479] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.589214] env[65788]: DEBUG nova.compute.manager [-] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 645.589387] env[65788]: DEBUG nova.network.neutron [-] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 645.589620] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.590219] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.590484] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.597727] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b3fe471-e063-4d04-8c99-9bc69c7b9e69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.612257] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661840, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.623885] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661841, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.625886] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 645.625886] env[65788]: value = "task-4661843" [ 645.625886] env[65788]: _type = "Task" [ 645.625886] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.635713] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.673561] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661842, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.716445] env[65788]: WARNING neutronclient.v2_0.client [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 645.717926] env[65788]: WARNING openstack [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.717926] env[65788]: WARNING openstack [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 645.810457] env[65788]: DEBUG nova.network.neutron [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Successfully created port: 4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 645.815237] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 646.010022] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 646.075641] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661840, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.135573] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661841, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.773584} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.140023] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 646.140023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 646.140023] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5856d50-2042-4acd-afe7-2257f2a29dca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.143730] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661843, 'name': ReconfigVM_Task, 'duration_secs': 0.418015} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.146831] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 141aefe8-1b95-4963-854d-da79ddf143f7/141aefe8-1b95-4963-854d-da79ddf143f7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 646.147972] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72af3b8f-ead6-4ac6-a322-502b38b6faa7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.153073] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 646.153073] env[65788]: value = "task-4661844" [ 646.153073] env[65788]: _type = "Task" [ 646.153073] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.158734] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 646.158734] env[65788]: value = "task-4661845" [ 646.158734] env[65788]: _type = "Task" [ 646.158734] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.185175] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661844, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.197702] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661842, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.028452} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.202472] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 646.202472] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 646.202895] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661845, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.206568] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d1dd78d-2668-4c3a-8730-c8e7b5169ed1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.216545] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 646.216545] env[65788]: value = "task-4661846" [ 646.216545] env[65788]: _type = "Task" [ 646.216545] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.230355] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.335301] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9093c14b-2e9b-4b1f-b640-cffae5bf48ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.344762] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc51cd66-70af-448c-9b5f-46f66484228a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.380357] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2bb97f-512b-48a6-9a5b-e0b8c82e4e6f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.389835] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3381a98-3ea5-464d-a141-811c922c245d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.408112] env[65788]: DEBUG nova.compute.provider_tree [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 646.567775] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661840, 'name': Rename_Task, 'duration_secs': 1.207837} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.568109] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 646.568388] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73915fc4-d524-4f51-8d03-3ea0cef93f99 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.576735] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 646.576735] env[65788]: value = "task-4661847" [ 646.576735] env[65788]: _type = "Task" [ 646.576735] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.593314] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661847, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.667909] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661844, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079062} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.668711] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 646.669583] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e99dcbf-90a9-49b3-9553-67f688819047 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.677080] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661845, 'name': Rename_Task, 'duration_secs': 0.175719} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.677846] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 646.678117] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f29bfd39-16ac-4f12-ba45-8f9f5269d3c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.705551] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 646.707383] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05e974ef-9936-4a01-9dd6-704648c2bbca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.725395] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 646.725395] env[65788]: value = "task-4661848" [ 646.725395] env[65788]: _type = "Task" [ 646.725395] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.739170] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 646.739170] env[65788]: value = "task-4661849" [ 646.739170] env[65788]: _type = "Task" [ 646.739170] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.751565] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661848, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.751925] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079281} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.753609] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 646.754822] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e7a58f-877c-491a-9d4c-4f8d38ce5361 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.762764] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.782080] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 646.782958] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10f17694-c7a3-40f5-aedf-cbbc55d5b116 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.805098] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 646.805098] env[65788]: value = "task-4661850" [ 646.805098] env[65788]: _type = "Task" [ 646.805098] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.819545] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661850, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.832781] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 646.867318] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 646.867589] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 646.867738] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 646.867969] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 646.868127] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 646.868221] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 646.868423] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.868579] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 646.868784] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 646.868926] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 646.869081] env[65788]: DEBUG nova.virt.hardware [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 646.870098] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23c8f5a-eae7-4175-aca1-beb7caa93923 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.879358] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2803c03-9378-478a-a2b4-14321fce1c1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.949101] env[65788]: DEBUG nova.scheduler.client.report [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 33 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 646.950539] env[65788]: DEBUG nova.compute.provider_tree [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 33 to 34 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 646.950539] env[65788]: DEBUG nova.compute.provider_tree [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 647.087473] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661847, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.100374] env[65788]: WARNING openstack [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.100771] env[65788]: WARNING openstack [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.243855] env[65788]: DEBUG oslo_vmware.api [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661848, 'name': PowerOnVM_Task, 'duration_secs': 0.504697} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.246973] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.247205] env[65788]: DEBUG nova.compute.manager [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 647.248016] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fd0120-b6cd-4e3e-b185-f89d56ef90ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.257771] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.313499] env[65788]: DEBUG nova.network.neutron [-] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 647.320487] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661850, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.456370] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.664s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 647.457138] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 647.460428] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.377s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.463055] env[65788]: INFO nova.compute.claims [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.590554] env[65788]: DEBUG oslo_vmware.api [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661847, 'name': PowerOnVM_Task, 'duration_secs': 0.997402} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.591047] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.591382] env[65788]: INFO nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Took 10.85 seconds to spawn the instance on the hypervisor. [ 647.591382] env[65788]: DEBUG nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 647.592636] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9138dc-a4ee-4d23-a568-82eb39d2ec27 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.613528] env[65788]: DEBUG nova.network.neutron [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Successfully updated port: 4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 647.687741] env[65788]: WARNING neutronclient.v2_0.client [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.688460] env[65788]: WARNING openstack [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.688818] env[65788]: WARNING openstack [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 647.753885] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661849, 'name': ReconfigVM_Task, 'duration_secs': 0.74465} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.754882] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 647.756661] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a83c6ba2-a3a7-4416-bda7-acb76d93c60c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.770020] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 647.770020] env[65788]: value = "task-4661851" [ 647.770020] env[65788]: _type = "Task" [ 647.770020] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.779390] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.787550] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661851, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.816941] env[65788]: INFO nova.compute.manager [-] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Took 2.23 seconds to deallocate network for instance. [ 647.822337] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661850, 'name': ReconfigVM_Task, 'duration_secs': 0.778623} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.822622] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae/54b0945e-c9f5-4726-928e-de1d0ded5aae.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 647.824113] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-956d6bce-a9d1-4619-89e6-f7173ea28fa8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.836273] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 647.836273] env[65788]: value = "task-4661852" [ 647.836273] env[65788]: _type = "Task" [ 647.836273] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.848304] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661852, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.861322] env[65788]: DEBUG nova.network.neutron [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Updated VIF entry in instance network info cache for port ca17fd35-e868-4f50-ac40-b6da3a12b015. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 647.861747] env[65788]: DEBUG nova.network.neutron [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Updating instance_info_cache with network_info: [{"id": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "address": "fa:16:3e:7e:b8:93", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca17fd35-e8", "ovs_interfaceid": "ca17fd35-e868-4f50-ac40-b6da3a12b015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 647.973395] env[65788]: DEBUG nova.compute.utils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 647.978151] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 647.978388] env[65788]: DEBUG nova.network.neutron [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 647.978730] env[65788]: WARNING neutronclient.v2_0.client [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.979621] env[65788]: WARNING neutronclient.v2_0.client [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 647.980961] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.981411] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.075446] env[65788]: DEBUG nova.policy [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '193d1b4346bf4d40bd184cc729e057ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca3090d0ab424469aa3fa576736bca1b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 648.115253] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "refresh_cache-d3e2b5f6-20f1-40ac-b92e-e110ee119178" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.116114] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "refresh_cache-d3e2b5f6-20f1-40ac-b92e-e110ee119178" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.116114] env[65788]: DEBUG nova.network.neutron [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 648.117617] env[65788]: INFO nova.compute.manager [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Took 21.96 seconds to build instance. [ 648.285299] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661851, 'name': Rename_Task, 'duration_secs': 0.231564} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.286027] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 648.286027] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1850afc2-64a4-4c02-a695-be053bfbda96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.294234] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 648.294234] env[65788]: value = "task-4661853" [ 648.294234] env[65788]: _type = "Task" [ 648.294234] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.303907] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.332134] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.352323] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661852, 'name': Rename_Task, 'duration_secs': 0.247259} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.352703] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 648.353176] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bad3daa9-8215-4c68-906f-20c4e6fe5d51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.363691] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 648.363691] env[65788]: value = "task-4661854" [ 648.363691] env[65788]: _type = "Task" [ 648.363691] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.370754] env[65788]: DEBUG oslo_concurrency.lockutils [req-251daf6b-c30f-4ae4-a7b3-7488fce4a15d req-be65e87d-bcab-4ad9-a80f-f5becb228fd0 service nova] Releasing lock "refresh_cache-77dcec96-50e5-4753-b3cb-c6aec377bce2" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 648.377509] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661854, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.477310] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 648.519383] env[65788]: DEBUG nova.network.neutron [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Successfully created port: 6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 648.578538] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.578783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.619229] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.619630] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.630275] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5d6595f3-ee4e-4140-b635-fd37deae1481 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "15aba949-c04c-4021-add8-2e9d4e4771db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.488s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.808559] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661853, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.875366] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661854, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.899071] env[65788]: DEBUG nova.network.neutron [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 648.928597] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4145407e-0e91-44b5-a1fa-de30879092de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.938216] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aeacf8f-db1b-46ae-a035-f51442ae0bd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.978105] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.979046] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.986016] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91ca92b-8e34-4fcb-a8d8-58105b348dc9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.998611] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d563946-f736-4fc9-b2bf-df4d6078d75c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.014167] env[65788]: DEBUG nova.compute.provider_tree [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.036489] env[65788]: WARNING neutronclient.v2_0.client [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 649.037158] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.037491] env[65788]: WARNING openstack [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 649.081111] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 649.161974] env[65788]: DEBUG nova.network.neutron [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Updating instance_info_cache with network_info: [{"id": "4a2afbad-2811-48fa-aee9-aeefd31b60d1", "address": "fa:16:3e:e1:2c:28", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2afbad-28", "ovs_interfaceid": "4a2afbad-2811-48fa-aee9-aeefd31b60d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.285123] env[65788]: DEBUG nova.compute.manager [req-b3aaceaa-6711-48bd-9e9d-cedd53d8e543 req-43b8abbb-7b23-47c7-9095-6649cbbd9d84 service nova] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Received event network-vif-deleted-c4fe2c1f-0ceb-4270-9ae5-7aa0b74d2b8b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 649.307189] env[65788]: DEBUG oslo_vmware.api [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661853, 'name': PowerOnVM_Task, 'duration_secs': 0.656566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.308030] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 649.308198] env[65788]: INFO nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Took 8.32 seconds to spawn the instance on the hypervisor. [ 649.308439] env[65788]: DEBUG nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 649.309617] env[65788]: DEBUG nova.compute.manager [req-988f5f1c-7bc5-412c-8221-035ed18eb162 req-63877876-b4e6-4240-b4f9-9d98da79a518 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Received event network-vif-plugged-4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 649.309815] env[65788]: DEBUG oslo_concurrency.lockutils [req-988f5f1c-7bc5-412c-8221-035ed18eb162 req-63877876-b4e6-4240-b4f9-9d98da79a518 service nova] Acquiring lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.310106] env[65788]: DEBUG oslo_concurrency.lockutils [req-988f5f1c-7bc5-412c-8221-035ed18eb162 req-63877876-b4e6-4240-b4f9-9d98da79a518 service nova] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.310389] env[65788]: DEBUG oslo_concurrency.lockutils [req-988f5f1c-7bc5-412c-8221-035ed18eb162 req-63877876-b4e6-4240-b4f9-9d98da79a518 service nova] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.310489] env[65788]: DEBUG nova.compute.manager [req-988f5f1c-7bc5-412c-8221-035ed18eb162 req-63877876-b4e6-4240-b4f9-9d98da79a518 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] No waiting events found dispatching network-vif-plugged-4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 649.310649] env[65788]: WARNING nova.compute.manager [req-988f5f1c-7bc5-412c-8221-035ed18eb162 req-63877876-b4e6-4240-b4f9-9d98da79a518 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Received unexpected event network-vif-plugged-4a2afbad-2811-48fa-aee9-aeefd31b60d1 for instance with vm_state building and task_state spawning. [ 649.311615] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1347ee7d-0ce8-4493-ad72-183264311c29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.374904] env[65788]: DEBUG oslo_vmware.api [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661854, 'name': PowerOnVM_Task, 'duration_secs': 0.621053} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.375505] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 649.375505] env[65788]: DEBUG nova.compute.manager [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 649.377062] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6211b21d-e01e-4d17-8d16-fc62e702ecec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.493703] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 649.518591] env[65788]: DEBUG nova.scheduler.client.report [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 649.524117] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 649.524423] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 649.524579] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 649.524754] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 649.524893] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 649.525044] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 649.525629] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.525832] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 649.526011] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 649.526270] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 649.526498] env[65788]: DEBUG nova.virt.hardware [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 649.527789] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9359f163-9135-4e8c-b0e4-16c836e220da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.540026] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5deaf2f7-3590-4713-82c0-3a88e6d06a0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.613744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.665687] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "refresh_cache-d3e2b5f6-20f1-40ac-b92e-e110ee119178" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 649.666369] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Instance network_info: |[{"id": "4a2afbad-2811-48fa-aee9-aeefd31b60d1", "address": "fa:16:3e:e1:2c:28", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2afbad-28", "ovs_interfaceid": "4a2afbad-2811-48fa-aee9-aeefd31b60d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 649.666636] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:2c:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a2afbad-2811-48fa-aee9-aeefd31b60d1', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 649.675706] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 649.676084] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 649.676391] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f861a02c-7dae-4c13-96fc-efbcb7fff213 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.700682] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 649.700682] env[65788]: value = "task-4661855" [ 649.700682] env[65788]: _type = "Task" [ 649.700682] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.710466] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661855, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.832764] env[65788]: INFO nova.compute.manager [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Took 19.36 seconds to build instance. [ 649.898118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.033949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.034415] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 650.037226] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.650s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.037444] env[65788]: DEBUG nova.objects.instance [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lazy-loading 'resources' on Instance uuid 4bedb939-e86b-42bd-a490-a73086ecfd7d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 650.199377] env[65788]: DEBUG nova.network.neutron [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Successfully updated port: 6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 650.218559] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661855, 'name': CreateVM_Task, 'duration_secs': 0.422376} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.219201] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 650.220379] env[65788]: WARNING neutronclient.v2_0.client [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.220784] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.220932] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.221336] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 650.221907] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df213025-f4a1-4da1-af80-dd7ddeb2e770 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.229021] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 650.229021] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529aefc5-396b-8dbd-ce00-e4379ee8130d" [ 650.229021] env[65788]: _type = "Task" [ 650.229021] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.242491] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529aefc5-396b-8dbd-ce00-e4379ee8130d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.335362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-540a916f-e592-4152-b90d-e836480ee626 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.875s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.541282] env[65788]: DEBUG nova.compute.utils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 650.545937] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 650.546156] env[65788]: DEBUG nova.network.neutron [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 650.546502] env[65788]: WARNING neutronclient.v2_0.client [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.546747] env[65788]: WARNING neutronclient.v2_0.client [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 650.547414] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.548482] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.702983] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.703310] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquired lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.706473] env[65788]: DEBUG nova.network.neutron [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 650.745657] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529aefc5-396b-8dbd-ce00-e4379ee8130d, 'name': SearchDatastore_Task, 'duration_secs': 0.012833} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.746358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.746358] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 650.746508] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.746663] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.746737] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 650.746985] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-669a88eb-71ba-4b93-bbf0-1747bc301bb7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.761468] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 650.761680] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 650.762524] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd71b5c0-4cc5-4196-b5f0-8eb18feef090 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.769154] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 650.769154] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52db8f1f-5226-2e11-c13a-75adf5b60d0b" [ 650.769154] env[65788]: _type = "Task" [ 650.769154] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.778689] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52db8f1f-5226-2e11-c13a-75adf5b60d0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.946823] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec9bd6c-952e-4f82-80c4-df8362a1d027 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.956444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e58c22-e663-4e80-a9c0-3aec5d57c6fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.994979] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2087f5f5-4676-4d48-b4fe-ae23b9c8a75e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.004429] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a279150-b5d6-4dd1-ba9e-d8313412c39a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.021445] env[65788]: DEBUG nova.compute.provider_tree [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.046451] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 651.212999] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.213456] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.282561] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52db8f1f-5226-2e11-c13a-75adf5b60d0b, 'name': SearchDatastore_Task, 'duration_secs': 0.025734} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.283529] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2650189d-05ae-4f40-aecb-3ff36be834a1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.292340] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 651.292340] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52912757-89dd-cf25-0cb7-a52ed0fb2a25" [ 651.292340] env[65788]: _type = "Task" [ 651.292340] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.303905] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52912757-89dd-cf25-0cb7-a52ed0fb2a25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.526210] env[65788]: DEBUG nova.scheduler.client.report [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 651.752781] env[65788]: DEBUG nova.network.neutron [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 651.809533] env[65788]: DEBUG nova.policy [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a7ca4f0daa04471b948f6affd617df3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1652d71008d45b28f4139d6edd8bf2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 651.812486] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52912757-89dd-cf25-0cb7-a52ed0fb2a25, 'name': SearchDatastore_Task, 'duration_secs': 0.028127} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.813346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.813663] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d3e2b5f6-20f1-40ac-b92e-e110ee119178/d3e2b5f6-20f1-40ac-b92e-e110ee119178.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 651.814186] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f61edcbc-72f8-41ce-9255-60e2e2787b30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.824537] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 651.824537] env[65788]: value = "task-4661856" [ 651.824537] env[65788]: _type = "Task" [ 651.824537] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.837868] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661856, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.034941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.998s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.037688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.622s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 652.039920] env[65788]: INFO nova.compute.claims [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.058338] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 652.071039] env[65788]: INFO nova.scheduler.client.report [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Deleted allocations for instance 4bedb939-e86b-42bd-a490-a73086ecfd7d [ 652.090568] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 652.091424] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 652.091773] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 652.092037] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 652.092335] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 652.092452] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 652.092805] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.093032] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 652.093237] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 652.093467] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 652.093667] env[65788]: DEBUG nova.virt.hardware [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 652.095190] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7580b1fa-4202-4a44-aac6-2a2a3b9c2705 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.106818] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e94e1e-bdbc-47e0-9041-02f86dc0e955 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.238351] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.238790] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.338745] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661856, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.581184] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1b8e67e8-0b2f-459c-ad4c-703cc738a589 tempest-ServerExternalEventsTest-1308199313 tempest-ServerExternalEventsTest-1308199313-project-member] Lock "4bedb939-e86b-42bd-a490-a73086ecfd7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.289s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.632412] env[65788]: DEBUG nova.network.neutron [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Successfully created port: b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 652.838167] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661856, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.868759} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.838167] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d3e2b5f6-20f1-40ac-b92e-e110ee119178/d3e2b5f6-20f1-40ac-b92e-e110ee119178.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 652.838167] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.838167] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-abf5e677-6680-4523-9e5c-5a54c60ac2b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.846488] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 652.846488] env[65788]: value = "task-4661857" [ 652.846488] env[65788]: _type = "Task" [ 652.846488] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.860662] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.360477] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.285363} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.363924] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 653.365056] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a787092b-fc33-4dcc-a478-db366f40dacf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.390105] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] d3e2b5f6-20f1-40ac-b92e-e110ee119178/d3e2b5f6-20f1-40ac-b92e-e110ee119178.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 653.393499] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63e89585-cf0c-4804-8d15-fdfb9ce40ebd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.410505] env[65788]: WARNING neutronclient.v2_0.client [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 653.411302] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.411903] env[65788]: WARNING openstack [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.428324] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 653.428324] env[65788]: value = "task-4661858" [ 653.428324] env[65788]: _type = "Task" [ 653.428324] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.438225] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661858, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.548230] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bcd8c2-d999-45be-9b7f-1f856f7155f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.558053] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934574fe-4afe-45dd-90a6-d402698b6139 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.602271] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08666a81-0d27-425d-a154-6e802dd99d6d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.617501] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3a38f5-be0b-43b1-ace9-18333cf8a687 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.634523] env[65788]: DEBUG nova.compute.provider_tree [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.663652] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "ed55713e-8bf1-4960-8309-47498b163d02" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.664170] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 653.664170] env[65788]: DEBUG nova.compute.manager [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Going to confirm migration 1 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 653.847510] env[65788]: DEBUG nova.network.neutron [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updating instance_info_cache with network_info: [{"id": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "address": "fa:16:3e:33:55:aa", "network": {"id": "62f163be-5383-479b-82ff-90e2dec87f17", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1288564708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca3090d0ab424469aa3fa576736bca1b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6209b956-7c", "ovs_interfaceid": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 653.940788] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661858, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.141030] env[65788]: DEBUG nova.scheduler.client.report [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 654.170857] env[65788]: WARNING neutronclient.v2_0.client [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.323921] env[65788]: WARNING neutronclient.v2_0.client [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 654.324618] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.324909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.325114] env[65788]: DEBUG nova.network.neutron [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 654.325894] env[65788]: DEBUG nova.objects.instance [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lazy-loading 'info_cache' on Instance uuid ed55713e-8bf1-4960-8309-47498b163d02 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 654.353101] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Releasing lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 654.353101] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Instance network_info: |[{"id": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "address": "fa:16:3e:33:55:aa", "network": {"id": "62f163be-5383-479b-82ff-90e2dec87f17", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1288564708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca3090d0ab424469aa3fa576736bca1b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6209b956-7c", "ovs_interfaceid": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 654.353310] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:55:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4345ef6-a7c8-4c1c-badf-a0d4f578b61c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6209b956-7cf9-4f9d-b25d-a36cca24051a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 654.361438] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Creating folder: Project (ca3090d0ab424469aa3fa576736bca1b). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 654.361803] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0963a343-e3ec-40db-bd17-e253fe77fd1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.377014] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Created folder: Project (ca3090d0ab424469aa3fa576736bca1b) in parent group-v910111. [ 654.377303] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Creating folder: Instances. Parent ref: group-v910167. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 654.378230] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d649d27b-15b2-4441-ada4-d3241ca72332 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.390743] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Created folder: Instances in parent group-v910167. [ 654.391559] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 654.391737] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 654.392051] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5cc455cd-dc98-46c1-91e6-3022d277713e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.424090] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 654.424090] env[65788]: value = "task-4661861" [ 654.424090] env[65788]: _type = "Task" [ 654.424090] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.440591] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661861, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.448826] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661858, 'name': ReconfigVM_Task, 'duration_secs': 0.618055} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.449153] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Reconfigured VM instance instance-00000013 to attach disk [datastore1] d3e2b5f6-20f1-40ac-b92e-e110ee119178/d3e2b5f6-20f1-40ac-b92e-e110ee119178.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 654.450198] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d29e64ca-0e56-4245-93c5-80ab8e67b122 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.458847] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 654.458847] env[65788]: value = "task-4661862" [ 654.458847] env[65788]: _type = "Task" [ 654.458847] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.468681] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661862, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.569134] env[65788]: DEBUG nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Received event network-changed-4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 654.569341] env[65788]: DEBUG nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Refreshing instance network info cache due to event network-changed-4a2afbad-2811-48fa-aee9-aeefd31b60d1. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 654.569752] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Acquiring lock "refresh_cache-d3e2b5f6-20f1-40ac-b92e-e110ee119178" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.570796] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Acquired lock "refresh_cache-d3e2b5f6-20f1-40ac-b92e-e110ee119178" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.571014] env[65788]: DEBUG nova.network.neutron [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Refreshing network info cache for port 4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 654.646880] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 654.648318] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 654.651230] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.519s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.653501] env[65788]: INFO nova.compute.claims [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 654.667298] env[65788]: DEBUG nova.network.neutron [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Successfully updated port: b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 654.938025] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661861, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.970318] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661862, 'name': Rename_Task, 'duration_secs': 0.229852} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.970592] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 654.970864] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a73afdb6-0a6a-439d-affb-38d4529f4537 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.982028] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 654.982028] env[65788]: value = "task-4661863" [ 654.982028] env[65788]: _type = "Task" [ 654.982028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.990217] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.073704] env[65788]: WARNING neutronclient.v2_0.client [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.074387] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.074749] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.161786] env[65788]: DEBUG nova.compute.utils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 655.164176] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 655.164495] env[65788]: DEBUG nova.network.neutron [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 655.164913] env[65788]: WARNING neutronclient.v2_0.client [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.166011] env[65788]: WARNING neutronclient.v2_0.client [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.166327] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.166789] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.186928] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "141aefe8-1b95-4963-854d-da79ddf143f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.187178] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "141aefe8-1b95-4963-854d-da79ddf143f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.187650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "141aefe8-1b95-4963-854d-da79ddf143f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.188113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "141aefe8-1b95-4963-854d-da79ddf143f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.188363] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "141aefe8-1b95-4963-854d-da79ddf143f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.190244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "refresh_cache-27ce0779-2fa2-44d2-a424-072183c022a6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.190399] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired lock "refresh_cache-27ce0779-2fa2-44d2-a424-072183c022a6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.190570] env[65788]: DEBUG nova.network.neutron [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 655.191869] env[65788]: INFO nova.compute.manager [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Terminating instance [ 655.312437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "15aba949-c04c-4021-add8-2e9d4e4771db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.312728] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "15aba949-c04c-4021-add8-2e9d4e4771db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.312950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "15aba949-c04c-4021-add8-2e9d4e4771db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.313147] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "15aba949-c04c-4021-add8-2e9d4e4771db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.313315] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "15aba949-c04c-4021-add8-2e9d4e4771db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.316467] env[65788]: INFO nova.compute.manager [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Terminating instance [ 655.336062] env[65788]: WARNING neutronclient.v2_0.client [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.336351] env[65788]: WARNING openstack [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.336686] env[65788]: WARNING openstack [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.436224] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661861, 'name': CreateVM_Task, 'duration_secs': 0.655773} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.436579] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 655.436989] env[65788]: WARNING neutronclient.v2_0.client [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.437385] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.437541] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.437861] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 655.438243] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-557777c9-7b0b-44a6-9bf2-eb333b7f4a49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.447849] env[65788]: DEBUG nova.policy [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f57cdfdf8c344604a7aa42a0fad72190', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c59a93f8dfbd439aa17af14eb6b7e988', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 655.449996] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 655.449996] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524c153d-80ad-b91d-27fb-790c0aceb044" [ 655.449996] env[65788]: _type = "Task" [ 655.449996] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.460050] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524c153d-80ad-b91d-27fb-790c0aceb044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.490769] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661863, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.495445] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "54b0945e-c9f5-4726-928e-de1d0ded5aae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.495719] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "54b0945e-c9f5-4726-928e-de1d0ded5aae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.495988] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "54b0945e-c9f5-4726-928e-de1d0ded5aae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.496276] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "54b0945e-c9f5-4726-928e-de1d0ded5aae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.496615] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "54b0945e-c9f5-4726-928e-de1d0ded5aae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.500301] env[65788]: INFO nova.compute.manager [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Terminating instance [ 655.685898] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 655.696561] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.696888] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.706537] env[65788]: DEBUG nova.compute.manager [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 655.706537] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.707017] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d80613-b0c1-467f-935b-644ad7d03d3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.717571] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 655.717571] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee148cd7-959e-4b54-9cb3-a411241d5f07 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.731388] env[65788]: DEBUG oslo_vmware.api [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 655.731388] env[65788]: value = "task-4661864" [ 655.731388] env[65788]: _type = "Task" [ 655.731388] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.742783] env[65788]: DEBUG oslo_vmware.api [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.821394] env[65788]: DEBUG nova.compute.manager [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 655.821759] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.823170] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f345cc-1938-49aa-8688-a6868d1a0f59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.839547] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 655.839968] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e518d65-e40d-45d5-9d51-9ea36af702e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.852321] env[65788]: DEBUG oslo_vmware.api [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 655.852321] env[65788]: value = "task-4661865" [ 655.852321] env[65788]: _type = "Task" [ 655.852321] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.862382] env[65788]: DEBUG oslo_vmware.api [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661865, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.961647] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524c153d-80ad-b91d-27fb-790c0aceb044, 'name': SearchDatastore_Task, 'duration_secs': 0.022707} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.961942] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.962215] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 655.962485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.962843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.962938] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 655.965897] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6bd90e1-3f4b-445e-b41d-6b35257202f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.977947] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 655.978760] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 655.979334] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff2c89f3-5e8c-4016-b97a-6505754faa81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.989434] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 655.989434] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5266ec92-739c-b4c6-f3bd-b319cffb62d4" [ 655.989434] env[65788]: _type = "Task" [ 655.989434] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.997392] env[65788]: DEBUG oslo_vmware.api [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4661863, 'name': PowerOnVM_Task, 'duration_secs': 0.796995} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.998173] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 655.998497] env[65788]: INFO nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Took 9.17 seconds to spawn the instance on the hypervisor. [ 655.998736] env[65788]: DEBUG nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 655.999742] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecb2762-9d81-452a-8030-1f01fdd40d31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.007430] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "refresh_cache-54b0945e-c9f5-4726-928e-de1d0ded5aae" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.007621] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquired lock "refresh_cache-54b0945e-c9f5-4726-928e-de1d0ded5aae" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 656.007784] env[65788]: DEBUG nova.network.neutron [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 656.010208] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5266ec92-739c-b4c6-f3bd-b319cffb62d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.133677] env[65788]: DEBUG nova.network.neutron [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Successfully created port: 48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 656.136574] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fba674-5428-479a-8e15-42d4f9783672 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.145976] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e1f45f-ab29-46fb-9a79-13f78427035a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.180487] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09be8e8f-c52e-465f-9d0c-3e9e719c90e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.188696] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d29e00-e27e-4df9-8c4c-103b145d6da7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.207786] env[65788]: DEBUG nova.compute.provider_tree [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.243161] env[65788]: DEBUG oslo_vmware.api [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661864, 'name': PowerOffVM_Task, 'duration_secs': 0.222442} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.249076] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 656.249263] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 656.249972] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ef5b06a-3ffd-477a-a2f3-aab69b631b40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.272423] env[65788]: DEBUG nova.network.neutron [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 656.318835] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.319352] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.338548] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 656.338731] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 656.338857] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore2] 141aefe8-1b95-4963-854d-da79ddf143f7 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 656.339147] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-475eb4fe-993a-4c9c-8fd7-2fa4cf97d173 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.346980] env[65788]: DEBUG oslo_vmware.api [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 656.346980] env[65788]: value = "task-4661867" [ 656.346980] env[65788]: _type = "Task" [ 656.346980] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.364775] env[65788]: DEBUG oslo_vmware.api [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.377050] env[65788]: DEBUG oslo_vmware.api [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661865, 'name': PowerOffVM_Task, 'duration_secs': 0.493216} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.377369] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 656.377591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 656.377935] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7ec068d-4e8c-46d8-96ba-8d337fbc2146 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.459616] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 656.459858] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 656.460617] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Deleting the datastore file [datastore2] 15aba949-c04c-4021-add8-2e9d4e4771db {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 656.460740] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ad3579b-1205-49d8-a319-202a2c128b06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.470671] env[65788]: DEBUG oslo_vmware.api [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for the task: (returnval){ [ 656.470671] env[65788]: value = "task-4661869" [ 656.470671] env[65788]: _type = "Task" [ 656.470671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.481297] env[65788]: DEBUG oslo_vmware.api [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.501715] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5266ec92-739c-b4c6-f3bd-b319cffb62d4, 'name': SearchDatastore_Task, 'duration_secs': 0.02694} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.502299] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f6d1d4b-bba5-42d1-b5c7-75f5049fc29d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.512271] env[65788]: WARNING neutronclient.v2_0.client [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.512271] env[65788]: WARNING openstack [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.512423] env[65788]: WARNING openstack [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.529404] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 656.529404] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521585e9-c27c-fbe4-a2df-c531ae1bbc91" [ 656.529404] env[65788]: _type = "Task" [ 656.529404] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.539089] env[65788]: INFO nova.compute.manager [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Took 25.22 seconds to build instance. [ 656.549945] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521585e9-c27c-fbe4-a2df-c531ae1bbc91, 'name': SearchDatastore_Task, 'duration_secs': 0.021033} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.549945] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.549945] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 957f14fc-23f1-4334-b672-b3a75398f716/957f14fc-23f1-4334-b672-b3a75398f716.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 656.550417] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf1e3d03-a10b-4f36-9b64-ae35a5968acf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.559934] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 656.559934] env[65788]: value = "task-4661870" [ 656.559934] env[65788]: _type = "Task" [ 656.559934] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.573579] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.699424] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 656.713491] env[65788]: DEBUG nova.scheduler.client.report [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.742996] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 656.743392] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 656.743606] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 656.743925] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 656.744032] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 656.744183] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 656.744451] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.744780] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 656.744856] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 656.745058] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 656.745278] env[65788]: DEBUG nova.virt.hardware [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 656.746686] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fcf74f-bc96-49e6-a3d0-f2c969fc6980 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.756911] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf55dd63-a3e2-494a-a559-5cd7faef41d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.860998] env[65788]: DEBUG oslo_vmware.api [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275704} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.861306] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 656.861488] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 656.861671] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 656.861845] env[65788]: INFO nova.compute.manager [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 656.862116] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 656.862323] env[65788]: DEBUG nova.compute.manager [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 656.862439] env[65788]: DEBUG nova.network.neutron [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 656.862777] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.863351] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.863613] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 656.947839] env[65788]: INFO nova.compute.manager [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Rebuilding instance [ 656.988451] env[65788]: DEBUG oslo_vmware.api [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Task: {'id': task-4661869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257907} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.988734] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 656.988916] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 656.989186] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 656.989539] env[65788]: INFO nova.compute.manager [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Took 1.17 seconds to destroy the instance on the hypervisor. [ 656.989983] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 656.990407] env[65788]: DEBUG nova.compute.manager [-] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 656.990536] env[65788]: DEBUG nova.network.neutron [-] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 656.991032] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 656.991648] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 656.991915] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.007897] env[65788]: DEBUG nova.compute.manager [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 657.008850] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba360ab-5d7d-4788-9249-05133316645c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.043425] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d042b093-8960-4703-8ed6-2ba9ce2cda88 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.742s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.073417] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661870, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.222606] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.222606] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 657.224684] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.542s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.224963] env[65788]: DEBUG nova.objects.instance [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lazy-loading 'resources' on Instance uuid 39127362-8daf-43ff-99cb-ddb54c6fe095 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 657.574867] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.686307} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.581225] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 957f14fc-23f1-4334-b672-b3a75398f716/957f14fc-23f1-4334-b672-b3a75398f716.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 657.581225] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.581225] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e999c426-a9ca-4ae3-a7e4-bd0df8b4073d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.589688] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 657.589688] env[65788]: value = "task-4661871" [ 657.589688] env[65788]: _type = "Task" [ 657.589688] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.603696] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.731893] env[65788]: DEBUG nova.compute.utils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 657.734958] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 657.739019] env[65788]: DEBUG nova.network.neutron [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 657.739019] env[65788]: WARNING neutronclient.v2_0.client [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.739019] env[65788]: WARNING neutronclient.v2_0.client [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 657.739019] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.739019] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.960879] env[65788]: WARNING openstack [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.961390] env[65788]: WARNING openstack [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.973228] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.973542] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.033986] env[65788]: DEBUG nova.network.neutron [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 658.039483] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 658.043055] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1873d44-e91a-4a40-928f-1a2eb95ecfe4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.054065] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 658.054065] env[65788]: value = "task-4661872" [ 658.054065] env[65788]: _type = "Task" [ 658.054065] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.068776] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.101509] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088193} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.105237] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.108025] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb04b67a-3fc2-44fe-8965-7cccf8e1bfa1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.133463] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] 957f14fc-23f1-4334-b672-b3a75398f716/957f14fc-23f1-4334-b672-b3a75398f716.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.136739] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64330276-1864-476a-ae0f-89b9a78ad256 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.159273] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 658.159273] env[65788]: value = "task-4661873" [ 658.159273] env[65788]: _type = "Task" [ 658.159273] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.166751] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec39e58-e60e-4349-aad6-3662bbe6dd83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.177677] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ef8384-3bdc-4480-bdc8-9fa526bf2ef4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.181283] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661873, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.216373] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.219165] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def01d96-4da4-47e1-a173-5f4d80b13bf4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.229121] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edf346e-67c0-4086-aa0e-53161593c1ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.244765] env[65788]: DEBUG nova.compute.provider_tree [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.249843] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 658.526540] env[65788]: DEBUG nova.network.neutron [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Successfully updated port: 48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 658.560742] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 658.569757] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661872, 'name': PowerOffVM_Task, 'duration_secs': 0.177438} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.570857] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 658.570857] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 658.571387] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5061b0cd-fbbc-450b-bc44-95c740a90475 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.580612] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 658.580933] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85b76c45-eeb3-4a8d-9986-630b4803dfb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.609155] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 658.609549] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 658.609631] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleting the datastore file [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 658.613018] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3caab320-9990-469a-9a10-43e5705f731d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.617888] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 658.617888] env[65788]: value = "task-4661875" [ 658.617888] env[65788]: _type = "Task" [ 658.617888] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.628142] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.670405] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661873, 'name': ReconfigVM_Task, 'duration_secs': 0.325481} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.670580] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Reconfigured VM instance instance-00000014 to attach disk [datastore1] 957f14fc-23f1-4334-b672-b3a75398f716/957f14fc-23f1-4334-b672-b3a75398f716.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 658.671232] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f20435ae-c2d1-4a45-8e59-5bc31bbcc061 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.680017] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 658.680017] env[65788]: value = "task-4661876" [ 658.680017] env[65788]: _type = "Task" [ 658.680017] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.693362] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661876, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.750237] env[65788]: DEBUG nova.scheduler.client.report [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.030065] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "refresh_cache-2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.032481] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquired lock "refresh_cache-2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.032481] env[65788]: DEBUG nova.network.neutron [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 659.132170] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160477} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.132605] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 659.132874] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 659.133156] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.158588] env[65788]: DEBUG nova.policy [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a1e3cda6bf9439982465aec80977cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1713c7002a413fb27180469fded83e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 659.194553] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661876, 'name': Rename_Task, 'duration_secs': 0.169943} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.194938] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 659.195302] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0200a1d5-db27-451f-81de-d8066cddb537 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.204300] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 659.204300] env[65788]: value = "task-4661877" [ 659.204300] env[65788]: _type = "Task" [ 659.204300] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.214489] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.260019] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 659.262772] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.264923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.195s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.266375] env[65788]: INFO nova.compute.claims [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.303803] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 659.304073] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 659.304243] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 659.304436] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 659.304571] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 659.304712] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 659.304912] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.305222] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 659.305416] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 659.305590] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 659.305771] env[65788]: DEBUG nova.virt.hardware [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 659.306826] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7910a3-dec8-4d89-bd03-d691fe29919c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.312279] env[65788]: INFO nova.scheduler.client.report [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Deleted allocations for instance 39127362-8daf-43ff-99cb-ddb54c6fe095 [ 659.320589] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-563cd84f-5d86-4107-aa9c-9efabac95bd8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.539023] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.539023] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.547599] env[65788]: WARNING neutronclient.v2_0.client [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 659.548434] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.549790] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.721570] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661877, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.748632] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 659.749743] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 659.823315] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2e319db2-3635-4bd1-8170-2df7f2fc069e tempest-ServerDiagnosticsV248Test-2062439731 tempest-ServerDiagnosticsV248Test-2062439731-project-member] Lock "39127362-8daf-43ff-99cb-ddb54c6fe095" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.435s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 660.175749] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 660.176289] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 660.176523] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 660.176805] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 660.178510] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 660.178510] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 660.178510] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 660.178510] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 660.178510] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 660.178926] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 660.179073] env[65788]: DEBUG nova.virt.hardware [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 660.179962] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c186b2b-eace-480a-a1d0-292ad264b468 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.196213] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6ab114-9807-41f4-8570-4baec32ba9f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.212022] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 660.217958] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 660.223723] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 660.223880] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9135e9a9-c15f-43c0-aab5-dbbf27e2dc5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.248971] env[65788]: DEBUG oslo_vmware.api [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661877, 'name': PowerOnVM_Task, 'duration_secs': 0.622376} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.250668] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 660.250981] env[65788]: INFO nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Took 10.76 seconds to spawn the instance on the hypervisor. [ 660.251417] env[65788]: DEBUG nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 660.252103] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 660.252103] env[65788]: value = "task-4661878" [ 660.252103] env[65788]: _type = "Task" [ 660.252103] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.252507] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4232e751-0776-4944-a2f7-2473b1e9a675 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.270674] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.271656] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.271656] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.271656] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.271656] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.272520] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.272520] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 660.272520] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 660.287405] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661878, 'name': CreateVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.352639] env[65788]: WARNING neutronclient.v2_0.client [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.353348] env[65788]: WARNING openstack [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.353698] env[65788]: WARNING openstack [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.380953] env[65788]: WARNING neutronclient.v2_0.client [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.381510] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.382401] env[65788]: WARNING openstack [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.403132] env[65788]: DEBUG nova.network.neutron [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 660.458437] env[65788]: DEBUG nova.network.neutron [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Successfully created port: 290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 660.471705] env[65788]: DEBUG nova.network.neutron [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.495181] env[65788]: DEBUG nova.network.neutron [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.608190] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.608190] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.640966] env[65788]: DEBUG nova.network.neutron [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Updated VIF entry in instance network info cache for port 4a2afbad-2811-48fa-aee9-aeefd31b60d1. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 660.641390] env[65788]: DEBUG nova.network.neutron [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Updating instance_info_cache with network_info: [{"id": "4a2afbad-2811-48fa-aee9-aeefd31b60d1", "address": "fa:16:3e:e1:2c:28", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a2afbad-28", "ovs_interfaceid": "4a2afbad-2811-48fa-aee9-aeefd31b60d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.677586] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42912f83-2335-4d14-b92f-2e9f61ddbb48 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.687418] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac1b213-1910-47b0-ae14-4b652b1c4f0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.723694] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ce93ec-8401-43c5-a89f-9cdf0d2e0c3a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.733354] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8080c997-eced-4c34-8a19-21f07eaef0f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.748099] env[65788]: DEBUG nova.compute.provider_tree [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 660.767227] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661878, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.770123] env[65788]: DEBUG nova.network.neutron [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [{"id": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "address": "fa:16:3e:2c:93:04", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa53d50-ad", "ovs_interfaceid": "ffa53d50-ad0e-438c-994a-b0b7a97f85f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.771568] env[65788]: DEBUG nova.network.neutron [-] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.777110] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 660.794927] env[65788]: INFO nova.compute.manager [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Took 27.44 seconds to build instance. [ 660.821166] env[65788]: DEBUG nova.network.neutron [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Updating instance_info_cache with network_info: [{"id": "b6a744b9-f11d-4c57-8cbd-538c3c7d139c", "address": "fa:16:3e:c2:8c:83", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.197", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6a744b9-f1", "ovs_interfaceid": "b6a744b9-f11d-4c57-8cbd-538c3c7d139c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 660.976643] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Releasing lock "refresh_cache-54b0945e-c9f5-4726-928e-de1d0ded5aae" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.976643] env[65788]: DEBUG nova.compute.manager [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 660.976643] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.977214] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a37b2a-be3b-4091-b14d-73e75b1471ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.987029] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 660.987316] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-560c7d21-192f-4008-9439-01a40bde674c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.994850] env[65788]: DEBUG oslo_vmware.api [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 660.994850] env[65788]: value = "task-4661879" [ 660.994850] env[65788]: _type = "Task" [ 660.994850] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.998570] env[65788]: INFO nova.compute.manager [-] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Took 4.14 seconds to deallocate network for instance. [ 661.007114] env[65788]: DEBUG oslo_vmware.api [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.061494] env[65788]: WARNING neutronclient.v2_0.client [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.062246] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.062808] env[65788]: WARNING openstack [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.294780] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Releasing lock "refresh_cache-d3e2b5f6-20f1-40ac-b92e-e110ee119178" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.294780] env[65788]: DEBUG nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Received event network-vif-plugged-6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 661.294780] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Acquiring lock "957f14fc-23f1-4334-b672-b3a75398f716-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.294780] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Lock "957f14fc-23f1-4334-b672-b3a75398f716-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.294780] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Lock "957f14fc-23f1-4334-b672-b3a75398f716-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.300851] env[65788]: DEBUG nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] No waiting events found dispatching network-vif-plugged-6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 661.300851] env[65788]: WARNING nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Received unexpected event network-vif-plugged-6209b956-7cf9-4f9d-b25d-a36cca24051a for instance with vm_state building and task_state spawning. [ 661.300851] env[65788]: DEBUG nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Received event network-changed-6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 661.300851] env[65788]: DEBUG nova.compute.manager [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Refreshing instance network info cache due to event network-changed-6209b956-7cf9-4f9d-b25d-a36cca24051a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 661.300851] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Acquiring lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.301794] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Acquired lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.301794] env[65788]: DEBUG nova.network.neutron [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Refreshing network info cache for port 6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 661.301794] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661878, 'name': CreateVM_Task, 'duration_secs': 0.56571} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.301794] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 661.301794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.301794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.302059] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 661.302059] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de98604a-3024-4ddd-ae9e-a18fb5f66851 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.302059] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-ed55713e-8bf1-4960-8309-47498b163d02" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.302059] env[65788]: DEBUG nova.objects.instance [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lazy-loading 'migration_context' on Instance uuid ed55713e-8bf1-4960-8309-47498b163d02 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 661.302059] env[65788]: INFO nova.compute.manager [-] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Took 4.29 seconds to deallocate network for instance. [ 661.302059] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 661.302059] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e08298-668c-d8bd-f6fc-2ae610109786" [ 661.302059] env[65788]: _type = "Task" [ 661.302317] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.302317] env[65788]: ERROR nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [req-42e57c29-1736-4f10-b5f3-4e56efdd8689] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-42e57c29-1736-4f10-b5f3-4e56efdd8689"}]} [ 661.305362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-617256eb-2967-431a-88d8-e84594d93dc0 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "957f14fc-23f1-4334-b672-b3a75398f716" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.958s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.305499] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e08298-668c-d8bd-f6fc-2ae610109786, 'name': SearchDatastore_Task, 'duration_secs': 0.019836} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.305750] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.306440] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 661.306917] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.307084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 661.310687] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 661.310687] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe3ada9b-25fe-4d04-b470-0c8c73441fae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.315959] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 661.323026] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 661.323026] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 661.323026] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae971c02-6316-41ce-8931-4a9cf0e0e5b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.326027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Releasing lock "refresh_cache-27ce0779-2fa2-44d2-a424-072183c022a6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.326453] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Instance network_info: |[{"id": "b6a744b9-f11d-4c57-8cbd-538c3c7d139c", "address": "fa:16:3e:c2:8c:83", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.197", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6a744b9-f1", "ovs_interfaceid": "b6a744b9-f11d-4c57-8cbd-538c3c7d139c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 661.327376] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:8c:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6a744b9-f11d-4c57-8cbd-538c3c7d139c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 661.337541] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 661.341406] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 661.341603] env[65788]: DEBUG nova.compute.provider_tree [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 661.344376] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 661.345396] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d3cc76e-4170-40ce-aa15-662c988ab468 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.370024] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 661.373864] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 661.373864] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223c9a3-bb43-cf94-afd9-e834ca387c6d" [ 661.373864] env[65788]: _type = "Task" [ 661.373864] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.380755] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 661.380755] env[65788]: value = "task-4661880" [ 661.380755] env[65788]: _type = "Task" [ 661.380755] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.389747] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223c9a3-bb43-cf94-afd9-e834ca387c6d, 'name': SearchDatastore_Task, 'duration_secs': 0.020095} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.390869] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e2e7500-8a7c-41a8-9230-06fb91b1dcd7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.397314] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661880, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.400091] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 661.404304] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 661.404304] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522a383d-cf51-4aff-6061-e9027c9c81bb" [ 661.404304] env[65788]: _type = "Task" [ 661.404304] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.414000] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522a383d-cf51-4aff-6061-e9027c9c81bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.512733] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.512874] env[65788]: DEBUG oslo_vmware.api [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661879, 'name': PowerOffVM_Task, 'duration_secs': 0.190753} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.513039] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 661.513423] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 661.513522] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-75a349ef-cbf8-4a53-aff5-e8f707380712 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.540500] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 661.540500] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 661.540688] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Deleting the datastore file [datastore2] 54b0945e-c9f5-4726-928e-de1d0ded5aae {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 661.540934] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c23b54fd-8766-4a61-9e70-3708f0ace5f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.549891] env[65788]: DEBUG oslo_vmware.api [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for the task: (returnval){ [ 661.549891] env[65788]: value = "task-4661882" [ 661.549891] env[65788]: _type = "Task" [ 661.549891] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.561103] env[65788]: DEBUG oslo_vmware.api [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.578336] env[65788]: DEBUG nova.network.neutron [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Updating instance_info_cache with network_info: [{"id": "48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b", "address": "fa:16:3e:2e:c3:e9", "network": {"id": "31b2ec29-0807-4adc-90ac-7109ca91bf6a", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-691221391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c59a93f8dfbd439aa17af14eb6b7e988", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48e98881-cb", "ovs_interfaceid": "48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 661.653369] env[65788]: WARNING neutronclient.v2_0.client [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 661.654153] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 661.654841] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 661.781480] env[65788]: DEBUG nova.objects.base [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 661.782631] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b9cfbc-ff0a-48dd-9891-79071040409e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.803283] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.807519] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c17ca01e-ae28-447d-b48c-8793690a798e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.818470] env[65788]: DEBUG oslo_vmware.api [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 661.818470] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525508c4-e702-c791-fdd9-10e8f849dd17" [ 661.818470] env[65788]: _type = "Task" [ 661.818470] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.829864] env[65788]: DEBUG oslo_vmware.api [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525508c4-e702-c791-fdd9-10e8f849dd17, 'name': SearchDatastore_Task, 'duration_secs': 0.009625} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.830368] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.875545] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc1836c-9e19-4aa7-a15f-4db5f086b40e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.887305] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be52581-89bf-40d7-ace9-9a49ecc41dd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.897515] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661880, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.931801] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a65e03-fb27-4d34-816f-31a38af31a1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.944589] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.944810] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.946430] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522a383d-cf51-4aff-6061-e9027c9c81bb, 'name': SearchDatastore_Task, 'duration_secs': 0.014253} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.946533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.946872] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 661.947908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9466f066-51dc-4887-98d7-4b2b5136e2cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.952491] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bb24da9-f9be-455b-9f4f-657c5bdd2b02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.967544] env[65788]: DEBUG nova.compute.provider_tree [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 661.976024] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 661.976024] env[65788]: value = "task-4661883" [ 661.976024] env[65788]: _type = "Task" [ 661.976024] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.982281] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661883, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.059878] env[65788]: DEBUG oslo_vmware.api [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Task: {'id': task-4661882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205666} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.060218] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 662.060339] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 662.060637] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.060730] env[65788]: INFO nova.compute.manager [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Took 1.08 seconds to destroy the instance on the hypervisor. [ 662.060896] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 662.061120] env[65788]: DEBUG nova.compute.manager [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 662.061272] env[65788]: DEBUG nova.network.neutron [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 662.061532] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.062049] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.062332] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.081855] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Releasing lock "refresh_cache-2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 662.082438] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Instance network_info: |[{"id": "48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b", "address": "fa:16:3e:2e:c3:e9", "network": {"id": "31b2ec29-0807-4adc-90ac-7109ca91bf6a", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-691221391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c59a93f8dfbd439aa17af14eb6b7e988", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48e98881-cb", "ovs_interfaceid": "48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 662.082859] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:c3:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1cbd5e0e-9116-46f1-9748-13a73d2d7e75', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 662.090478] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Creating folder: Project (c59a93f8dfbd439aa17af14eb6b7e988). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 662.090773] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6df3e56e-9750-4574-9815-ad9aa22802bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.104628] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Created folder: Project (c59a93f8dfbd439aa17af14eb6b7e988) in parent group-v910111. [ 662.104829] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Creating folder: Instances. Parent ref: group-v910172. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 662.105149] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ab6e0fe-c66c-452c-9a84-957c84d509d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.118839] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Created folder: Instances in parent group-v910172. [ 662.119229] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 662.119558] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 662.119768] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11a28e6a-1da5-4766-a126-6684acf20e7c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.151237] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 662.151237] env[65788]: value = "task-4661886" [ 662.151237] env[65788]: _type = "Task" [ 662.151237] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.163830] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661886, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.239763] env[65788]: DEBUG nova.network.neutron [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Successfully updated port: 290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 662.291680] env[65788]: DEBUG nova.network.neutron [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 662.291979] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.394661] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 662.396034] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 662.410558] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661880, 'name': CreateVM_Task, 'duration_secs': 0.535956} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.411212] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 662.412794] env[65788]: WARNING neutronclient.v2_0.client [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 662.413490] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.413707] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 662.414285] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 662.414581] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f59e5c17-72b9-4e2b-aab3-5e61adefb8c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.421981] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 662.421981] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b64cac-6200-e344-1398-8276ee90eb63" [ 662.421981] env[65788]: _type = "Task" [ 662.421981] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.229369] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 663.234748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "refresh_cache-42a9c6f2-08f0-4548-9101-685fc3930c7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.234924] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "refresh_cache-42a9c6f2-08f0-4548-9101-685fc3930c7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.235097] env[65788]: DEBUG nova.network.neutron [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 663.236076] env[65788]: DEBUG nova.network.neutron [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.239312] env[65788]: DEBUG nova.compute.manager [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Received event network-vif-plugged-b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 663.239312] env[65788]: DEBUG oslo_concurrency.lockutils [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Acquiring lock "27ce0779-2fa2-44d2-a424-072183c022a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.239312] env[65788]: DEBUG oslo_concurrency.lockutils [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Lock "27ce0779-2fa2-44d2-a424-072183c022a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.239312] env[65788]: DEBUG oslo_concurrency.lockutils [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Lock "27ce0779-2fa2-44d2-a424-072183c022a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 663.239515] env[65788]: DEBUG nova.compute.manager [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] No waiting events found dispatching network-vif-plugged-b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 663.239591] env[65788]: WARNING nova.compute.manager [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Received unexpected event network-vif-plugged-b6a744b9-f11d-4c57-8cbd-538c3c7d139c for instance with vm_state building and task_state spawning. [ 663.239728] env[65788]: DEBUG nova.compute.manager [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Received event network-changed-b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 663.239863] env[65788]: DEBUG nova.compute.manager [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Refreshing instance network info cache due to event network-changed-b6a744b9-f11d-4c57-8cbd-538c3c7d139c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 663.240046] env[65788]: DEBUG oslo_concurrency.lockutils [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Acquiring lock "refresh_cache-27ce0779-2fa2-44d2-a424-072183c022a6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.242124] env[65788]: DEBUG oslo_concurrency.lockutils [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Acquired lock "refresh_cache-27ce0779-2fa2-44d2-a424-072183c022a6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.242124] env[65788]: DEBUG nova.network.neutron [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Refreshing network info cache for port b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 663.256942] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b64cac-6200-e344-1398-8276ee90eb63, 'name': SearchDatastore_Task, 'duration_secs': 0.023915} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.264153] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.264422] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.264726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.264818] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.265626] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.265626] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661886, 'name': CreateVM_Task, 'duration_secs': 0.742221} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.265765] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661883, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.779719} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.267131] env[65788]: ERROR nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [req-6c4e149c-b112-459f-ad56-c05e778d34d5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6c4e149c-b112-459f-ad56-c05e778d34d5"}]} [ 663.267929] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6e31a55-e0c4-4dd1-b7bb-2e05c76ff883 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.269902] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 663.271361] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 663.271361] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 663.273633] env[65788]: WARNING neutronclient.v2_0.client [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.274194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.274194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.274548] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 663.274783] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac57f521-5bb7-455d-9b91-3fc4ccce7e6f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.279527] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c90c149b-8320-4f95-a605-916f0783a42a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.287866] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 663.287866] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cc4268-2e9c-a95c-0d2b-dcd8e569c604" [ 663.287866] env[65788]: _type = "Task" [ 663.287866] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.287866] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 663.287866] env[65788]: value = "task-4661887" [ 663.287866] env[65788]: _type = "Task" [ 663.287866] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.289102] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.289278] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 663.290753] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 663.294034] env[65788]: WARNING neutronclient.v2_0.client [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.294649] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.295221] env[65788]: WARNING openstack [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.311567] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8f5e9c7-8601-4a36-936d-a002613cab5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.321142] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661887, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.328666] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 663.328666] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52feeaf1-c582-b52a-00fe-df81103dcf29" [ 663.328666] env[65788]: _type = "Task" [ 663.328666] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.328666] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cc4268-2e9c-a95c-0d2b-dcd8e569c604, 'name': SearchDatastore_Task, 'duration_secs': 0.015868} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.328666] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.328666] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 663.328920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.328920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.328920] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 663.331959] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c4cca95-3f46-4dc5-88cf-65ff67e0a02c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.337588] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 663.337819] env[65788]: DEBUG nova.compute.provider_tree [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 663.344402] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52feeaf1-c582-b52a-00fe-df81103dcf29, 'name': SearchDatastore_Task, 'duration_secs': 0.012164} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.347132] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 663.347132] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 663.347378] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef526ff9-655b-4228-9df7-b7eb81e346a1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.349645] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ccf9bc7-7d55-4b90-8bb2-aa43a13d40e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.356990] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 663.356990] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526f0524-92e9-2cbe-67fa-859a03b61b3d" [ 663.356990] env[65788]: _type = "Task" [ 663.356990] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.359485] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 663.359485] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5216a80a-fde2-ed0f-01e6-54ab239346f3" [ 663.359485] env[65788]: _type = "Task" [ 663.359485] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.363098] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 663.374061] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5216a80a-fde2-ed0f-01e6-54ab239346f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.377061] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526f0524-92e9-2cbe-67fa-859a03b61b3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.390634] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 663.564610] env[65788]: DEBUG nova.network.neutron [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updated VIF entry in instance network info cache for port 6209b956-7cf9-4f9d-b25d-a36cca24051a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 663.567281] env[65788]: DEBUG nova.network.neutron [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updating instance_info_cache with network_info: [{"id": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "address": "fa:16:3e:33:55:aa", "network": {"id": "62f163be-5383-479b-82ff-90e2dec87f17", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1288564708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca3090d0ab424469aa3fa576736bca1b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6209b956-7c", "ovs_interfaceid": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 663.744699] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.745153] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.784103] env[65788]: INFO nova.compute.manager [-] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Took 1.72 seconds to deallocate network for instance. [ 663.784770] env[65788]: WARNING neutronclient.v2_0.client [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 663.785438] env[65788]: WARNING openstack [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 663.785789] env[65788]: WARNING openstack [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 663.793620] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b92ba4-3abf-4310-a9ae-a490b9d71ba6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.797061] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.808837] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae69a82-8a2d-406e-b875-0d7112472cd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.815154] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661887, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082651} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.815890] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 663.816701] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba57796-3cc7-48d9-946e-32a8ad1678fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.845164] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa26705-b072-464e-b151-7ceec388396b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.864288] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 663.865536] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdfe0c75-46d3-48cc-840b-dbfe904a56e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.890197] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b0cba6-520c-4134-be60-f5d57a335db2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.896521] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 663.896521] env[65788]: value = "task-4661888" [ 663.896521] env[65788]: _type = "Task" [ 663.896521] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.907511] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5216a80a-fde2-ed0f-01e6-54ab239346f3, 'name': SearchDatastore_Task, 'duration_secs': 0.033404} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.907748] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526f0524-92e9-2cbe-67fa-859a03b61b3d, 'name': SearchDatastore_Task, 'duration_secs': 0.029746} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.910218] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.910524] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 27ce0779-2fa2-44d2-a424-072183c022a6/27ce0779-2fa2-44d2-a424-072183c022a6.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 663.911540] env[65788]: DEBUG nova.network.neutron [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 663.913587] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7a2e829-fe9b-4a9a-957b-04d09a797e84 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.915953] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5728b496-4f9a-4745-a552-d277d95e847b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.923935] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.932973] env[65788]: DEBUG nova.compute.provider_tree [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.936041] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 663.936041] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e579e-1817-f129-0f53-af90727347e7" [ 663.936041] env[65788]: _type = "Task" [ 663.936041] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.937786] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 663.937786] env[65788]: value = "task-4661889" [ 663.937786] env[65788]: _type = "Task" [ 663.937786] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.951909] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.955307] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e579e-1817-f129-0f53-af90727347e7, 'name': SearchDatastore_Task, 'duration_secs': 0.011741} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.955711] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 663.956164] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8/2529ef4c-80d8-47b1-a73a-a02eea0ad8d8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 663.956541] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-65d40a63-d257-41a7-a43b-5b243801a48c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.964817] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 663.964817] env[65788]: value = "task-4661890" [ 663.964817] env[65788]: _type = "Task" [ 663.964817] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.970739] env[65788]: DEBUG nova.compute.manager [req-7c92b692-9041-4570-a961-901343eabfa7 req-26afb3e7-3ac3-43e7-b158-73e5d13513e6 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Received event network-vif-plugged-290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 663.970937] env[65788]: DEBUG oslo_concurrency.lockutils [req-7c92b692-9041-4570-a961-901343eabfa7 req-26afb3e7-3ac3-43e7-b158-73e5d13513e6 service nova] Acquiring lock "42a9c6f2-08f0-4548-9101-685fc3930c7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 663.971291] env[65788]: DEBUG oslo_concurrency.lockutils [req-7c92b692-9041-4570-a961-901343eabfa7 req-26afb3e7-3ac3-43e7-b158-73e5d13513e6 service nova] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.971396] env[65788]: DEBUG oslo_concurrency.lockutils [req-7c92b692-9041-4570-a961-901343eabfa7 req-26afb3e7-3ac3-43e7-b158-73e5d13513e6 service nova] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 663.971535] env[65788]: DEBUG nova.compute.manager [req-7c92b692-9041-4570-a961-901343eabfa7 req-26afb3e7-3ac3-43e7-b158-73e5d13513e6 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] No waiting events found dispatching network-vif-plugged-290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 663.971718] env[65788]: WARNING nova.compute.manager [req-7c92b692-9041-4570-a961-901343eabfa7 req-26afb3e7-3ac3-43e7-b158-73e5d13513e6 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Received unexpected event network-vif-plugged-290d5ac2-d3c4-4f0d-b549-523d4ef21f1b for instance with vm_state building and task_state spawning. [ 663.978335] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661890, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.070845] env[65788]: DEBUG oslo_concurrency.lockutils [req-132fac7f-43f5-4efb-81b3-94a64c1f0e03 req-d018affd-c293-4960-9d87-fc9773f79c46 service nova] Releasing lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.281255] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.281945] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.309800] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 664.413103] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.439539] env[65788]: DEBUG nova.scheduler.client.report [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 664.462034] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661889, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.482699] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661890, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.744455] env[65788]: WARNING openstack [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.744850] env[65788]: WARNING openstack [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.803053] env[65788]: WARNING neutronclient.v2_0.client [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.803847] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.804250] env[65788]: WARNING openstack [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.837889] env[65788]: WARNING neutronclient.v2_0.client [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 664.837889] env[65788]: WARNING openstack [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.837889] env[65788]: WARNING openstack [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.904225] env[65788]: DEBUG nova.network.neutron [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Updating instance_info_cache with network_info: [{"id": "290d5ac2-d3c4-4f0d-b549-523d4ef21f1b", "address": "fa:16:3e:7c:a0:74", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap290d5ac2-d3", "ovs_interfaceid": "290d5ac2-d3c4-4f0d-b549-523d4ef21f1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 664.918958] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661888, 'name': ReconfigVM_Task, 'duration_secs': 0.735812} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.919228] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1/78b545b3-1f84-4ea8-b180-f9039f2ea4b1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 664.919982] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc1cab85-7638-4b36-a113-d0bd4c3a7c16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.929625] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 664.929625] env[65788]: value = "task-4661891" [ 664.929625] env[65788]: _type = "Task" [ 664.929625] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.939246] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661891, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.952032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.687s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.952619] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 664.964127] env[65788]: DEBUG nova.network.neutron [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Updated VIF entry in instance network info cache for port b6a744b9-f11d-4c57-8cbd-538c3c7d139c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 664.964330] env[65788]: DEBUG nova.network.neutron [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Updating instance_info_cache with network_info: [{"id": "b6a744b9-f11d-4c57-8cbd-538c3c7d139c", "address": "fa:16:3e:c2:8c:83", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.197", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6a744b9-f1", "ovs_interfaceid": "b6a744b9-f11d-4c57-8cbd-538c3c7d139c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 664.966619] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.295s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 664.968696] env[65788]: INFO nova.compute.claims [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.971659] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63685} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.976020] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 27ce0779-2fa2-44d2-a424-072183c022a6/27ce0779-2fa2-44d2-a424-072183c022a6.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 664.976020] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 664.977404] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0bf99696-ba1e-4dcb-82c8-fe0dcf8675df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.989345] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661890, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762526} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.992768] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8/2529ef4c-80d8-47b1-a73a-a02eea0ad8d8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 664.993035] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 664.993386] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 664.993386] env[65788]: value = "task-4661892" [ 664.993386] env[65788]: _type = "Task" [ 664.993386] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.995240] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a08e44c2-0815-422e-be48-5638bf6b1545 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.007947] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.009574] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 665.009574] env[65788]: value = "task-4661893" [ 665.009574] env[65788]: _type = "Task" [ 665.009574] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.022782] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661893, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.407715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "refresh_cache-42a9c6f2-08f0-4548-9101-685fc3930c7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.408466] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Instance network_info: |[{"id": "290d5ac2-d3c4-4f0d-b549-523d4ef21f1b", "address": "fa:16:3e:7c:a0:74", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap290d5ac2-d3", "ovs_interfaceid": "290d5ac2-d3c4-4f0d-b549-523d4ef21f1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 665.409014] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:a0:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '290d5ac2-d3c4-4f0d-b549-523d4ef21f1b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 665.417136] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating folder: Project (3e1713c7002a413fb27180469fded83e). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 665.417468] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd0be96c-6b9b-4331-808d-65fdbb7dc55f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.431430] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created folder: Project (3e1713c7002a413fb27180469fded83e) in parent group-v910111. [ 665.431914] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating folder: Instances. Parent ref: group-v910175. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 665.438110] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11a2bc65-5b7e-416d-aa54-5676ac4f14cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.446537] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661891, 'name': Rename_Task, 'duration_secs': 0.381511} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.446852] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 665.447065] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c92d13e-c778-4c26-87b8-626d098f91ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.450556] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created folder: Instances in parent group-v910175. [ 665.450728] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 665.450784] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 665.450997] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb01e329-a7bb-4ba8-8b42-75f4fa73acd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.467305] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 665.467305] env[65788]: value = "task-4661896" [ 665.467305] env[65788]: _type = "Task" [ 665.467305] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.473953] env[65788]: DEBUG nova.compute.utils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 665.475561] env[65788]: DEBUG oslo_concurrency.lockutils [req-72ec3fd5-f31f-4c48-b7a2-a4056ddb3bf6 req-303db5cf-9c1b-4c70-81bf-851d381c04f8 service nova] Releasing lock "refresh_cache-27ce0779-2fa2-44d2-a424-072183c022a6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 665.476739] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 665.476739] env[65788]: value = "task-4661897" [ 665.476739] env[65788]: _type = "Task" [ 665.476739] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.479927] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 665.480250] env[65788]: DEBUG nova.network.neutron [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 665.480789] env[65788]: WARNING neutronclient.v2_0.client [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.481156] env[65788]: WARNING neutronclient.v2_0.client [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 665.481813] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.482374] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.492870] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661896, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.504633] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661897, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.511855] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.189348} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.515203] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 665.516487] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4a27ef-81ad-4c4d-8f44-3333e5dfbd90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.545087] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] 27ce0779-2fa2-44d2-a424-072183c022a6/27ce0779-2fa2-44d2-a424-072183c022a6.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 665.549283] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec86ffc4-8613-40cc-99a7-f6bec5e60b18 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.563884] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661893, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177278} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.564881] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 665.566118] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2b9af8-8acd-404d-98c6-c57c70246971 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.571351] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 665.571351] env[65788]: value = "task-4661898" [ 665.571351] env[65788]: _type = "Task" [ 665.571351] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.593731] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8/2529ef4c-80d8-47b1-a73a-a02eea0ad8d8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 665.598933] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8571b79-c641-41d3-8e33-4666be6bb038 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.620171] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661898, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.621298] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 665.621298] env[65788]: value = "task-4661899" [ 665.621298] env[65788]: _type = "Task" [ 665.621298] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.630529] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.723327] env[65788]: DEBUG nova.policy [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e441ce452ee48ea8f4ab04c378be68d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89909e0b9fd9487982d43248f9022481', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 665.982430] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661896, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.995733] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 666.010972] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661897, 'name': CreateVM_Task, 'duration_secs': 0.496376} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.010972] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 666.010972] env[65788]: WARNING neutronclient.v2_0.client [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 666.010972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.010972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.010972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 666.011452] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d42f0fc-0149-4999-8b89-e0ccf323ae52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.020035] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 666.020035] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1bb92-519c-3858-e501-128740964425" [ 666.020035] env[65788]: _type = "Task" [ 666.020035] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.032895] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1bb92-519c-3858-e501-128740964425, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.095562] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661898, 'name': ReconfigVM_Task, 'duration_secs': 0.495849} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.095998] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Reconfigured VM instance instance-00000015 to attach disk [datastore1] 27ce0779-2fa2-44d2-a424-072183c022a6/27ce0779-2fa2-44d2-a424-072183c022a6.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 666.097042] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-092ebe80-165c-4a0d-98db-588a3b1fa5fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.105669] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 666.105669] env[65788]: value = "task-4661900" [ 666.105669] env[65788]: _type = "Task" [ 666.105669] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.116721] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661900, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.133263] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661899, 'name': ReconfigVM_Task, 'duration_secs': 0.355235} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.136590] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8/2529ef4c-80d8-47b1-a73a-a02eea0ad8d8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 666.137524] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac2ab7ef-f861-4cd5-8442-eee590a80d24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.145057] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 666.145057] env[65788]: value = "task-4661901" [ 666.145057] env[65788]: _type = "Task" [ 666.145057] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.156017] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661901, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.248398] env[65788]: DEBUG nova.network.neutron [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Successfully created port: 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 666.449810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e024af47-2a03-4533-9bcd-3d9551578c55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.459500] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044cafdd-41f3-49a3-aef7-f9bcfa92433c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.499230] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d08d7f-5f91-4f47-aefd-5d06b10499e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.512395] env[65788]: DEBUG oslo_vmware.api [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661896, 'name': PowerOnVM_Task, 'duration_secs': 0.548055} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.514043] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1fa944-0521-41e4-a482-f633694633d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.518359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 666.518625] env[65788]: DEBUG nova.compute.manager [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 666.519474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970ee377-5336-4a08-b018-fc2991762a57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.535171] env[65788]: DEBUG nova.compute.provider_tree [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 666.544931] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1bb92-519c-3858-e501-128740964425, 'name': SearchDatastore_Task, 'duration_secs': 0.014197} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.545555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.545702] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 666.545930] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.546084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.546262] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 666.546539] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f57ddcd-ff4d-4d91-ac97-50eb151dc43f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.558223] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 666.558223] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 666.558223] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-332a6311-a5db-4095-becc-7324e9cef95a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.565520] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 666.565520] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bd2ed4-4d19-d2cf-4638-fabc7f00e537" [ 666.565520] env[65788]: _type = "Task" [ 666.565520] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.574763] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bd2ed4-4d19-d2cf-4638-fabc7f00e537, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.616062] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661900, 'name': Rename_Task, 'duration_secs': 0.354839} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.616451] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 666.617757] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a12a52f-203f-4366-9d6f-f53328ff949a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.625764] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 666.625764] env[65788]: value = "task-4661906" [ 666.625764] env[65788]: _type = "Task" [ 666.625764] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.635130] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.656607] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661901, 'name': Rename_Task, 'duration_secs': 0.251969} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.657221] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 666.657221] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e600b2b3-8633-411d-af26-d0a495927e47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.665496] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 666.665496] env[65788]: value = "task-4661907" [ 666.665496] env[65788]: _type = "Task" [ 666.665496] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.674887] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.999429] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.999994] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.015014] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 667.036121] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 667.036457] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 667.036632] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 667.036835] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 667.036984] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 667.037127] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 667.037337] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.037505] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 667.037663] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 667.037821] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 667.038057] env[65788]: DEBUG nova.virt.hardware [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 667.039038] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f20ab4-8095-4019-9e40-71dd2c8c1928 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.058017] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf105b3e-5221-4356-9bad-a8b1a350bf7f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.062433] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.064263] env[65788]: ERROR nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [req-3eecd5c1-5b11-4d3f-8ffb-c776b7b7a2b5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3eecd5c1-5b11-4d3f-8ffb-c776b7b7a2b5"}]} [ 667.087197] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bd2ed4-4d19-d2cf-4638-fabc7f00e537, 'name': SearchDatastore_Task, 'duration_secs': 0.012111} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.088680] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76b59b70-8969-45d0-978a-2147301c3fb4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.092027] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 667.099273] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 667.099273] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5244192c-3cb8-234c-536a-1b6ad9d88354" [ 667.099273] env[65788]: _type = "Task" [ 667.099273] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.108290] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5244192c-3cb8-234c-536a-1b6ad9d88354, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.109784] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 667.109979] env[65788]: DEBUG nova.compute.provider_tree [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 667.122261] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 667.137857] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661906, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.143646] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 667.177896] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661907, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.347007] env[65788]: DEBUG nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Received event network-vif-plugged-48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 667.347345] env[65788]: DEBUG oslo_concurrency.lockutils [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Acquiring lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.347650] env[65788]: DEBUG oslo_concurrency.lockutils [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.347893] env[65788]: DEBUG oslo_concurrency.lockutils [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.348199] env[65788]: DEBUG nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] No waiting events found dispatching network-vif-plugged-48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 667.348434] env[65788]: WARNING nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Received unexpected event network-vif-plugged-48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b for instance with vm_state building and task_state spawning. [ 667.348697] env[65788]: DEBUG nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Received event network-changed-48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 667.348901] env[65788]: DEBUG nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Refreshing instance network info cache due to event network-changed-48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 667.349201] env[65788]: DEBUG oslo_concurrency.lockutils [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Acquiring lock "refresh_cache-2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.349399] env[65788]: DEBUG oslo_concurrency.lockutils [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Acquired lock "refresh_cache-2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.349609] env[65788]: DEBUG nova.network.neutron [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Refreshing network info cache for port 48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 667.498446] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd83dd52-89bd-4385-a4e5-7931ff0986a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.502637] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 667.517152] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aad0de7-a432-414d-9ab8-a509ef76e950 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.559660] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03dc44b-a3a6-4882-a730-d863328c5e86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.573302] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc39e86b-a4ad-4efd-9213-f362376345c6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.590567] env[65788]: DEBUG nova.compute.manager [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 667.591167] env[65788]: DEBUG nova.compute.provider_tree [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 667.593280] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d938b36-6251-4d06-b3f7-42f6f535609e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.613561] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5244192c-3cb8-234c-536a-1b6ad9d88354, 'name': SearchDatastore_Task, 'duration_secs': 0.020833} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.614197] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.614463] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 42a9c6f2-08f0-4548-9101-685fc3930c7d/42a9c6f2-08f0-4548-9101-685fc3930c7d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 667.614737] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c854f13b-e680-4669-8dc6-50749826039c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.625304] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 667.625304] env[65788]: value = "task-4661908" [ 667.625304] env[65788]: _type = "Task" [ 667.625304] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.638413] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661908, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.642996] env[65788]: DEBUG oslo_vmware.api [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661906, 'name': PowerOnVM_Task, 'duration_secs': 0.776583} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.643305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 667.643505] env[65788]: INFO nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Took 15.59 seconds to spawn the instance on the hypervisor. [ 667.643679] env[65788]: DEBUG nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 667.644984] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c1cf4a-0ef1-4b10-abf5-7e201c28025f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.678324] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661907, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.840325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "670f59c3-10b0-431f-a8b0-ef6f82a938be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.840613] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.853518] env[65788]: WARNING neutronclient.v2_0.client [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 667.856015] env[65788]: WARNING openstack [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.856015] env[65788]: WARNING openstack [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.979772] env[65788]: DEBUG nova.network.neutron [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Successfully updated port: 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 668.031148] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.048656] env[65788]: WARNING openstack [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.049089] env[65788]: WARNING openstack [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.107716] env[65788]: INFO nova.compute.manager [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] instance snapshotting [ 668.108440] env[65788]: DEBUG nova.objects.instance [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'flavor' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 668.126936] env[65788]: ERROR nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [req-bb07f4e7-bfee-4e33-87d7-d09703222084] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bb07f4e7-bfee-4e33-87d7-d09703222084"}]} [ 668.132618] env[65788]: WARNING neutronclient.v2_0.client [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.133279] env[65788]: WARNING openstack [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.133663] env[65788]: WARNING openstack [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.154939] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661908, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.157359] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 668.183853] env[65788]: INFO nova.compute.manager [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Took 34.12 seconds to build instance. [ 668.187729] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 668.187929] env[65788]: DEBUG nova.compute.provider_tree [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 668.195524] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661907, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.204381] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 668.228247] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 668.277299] env[65788]: DEBUG nova.network.neutron [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Updated VIF entry in instance network info cache for port 48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 668.277554] env[65788]: DEBUG nova.network.neutron [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Updating instance_info_cache with network_info: [{"id": "48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b", "address": "fa:16:3e:2e:c3:e9", "network": {"id": "31b2ec29-0807-4adc-90ac-7109ca91bf6a", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-691221391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c59a93f8dfbd439aa17af14eb6b7e988", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1cbd5e0e-9116-46f1-9748-13a73d2d7e75", "external-id": "nsx-vlan-transportzone-690", "segmentation_id": 690, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48e98881-cb", "ovs_interfaceid": "48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 668.343478] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 668.350362] env[65788]: DEBUG nova.compute.manager [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Received event network-changed-290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 668.350362] env[65788]: DEBUG nova.compute.manager [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Refreshing instance network info cache due to event network-changed-290d5ac2-d3c4-4f0d-b549-523d4ef21f1b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 668.350362] env[65788]: DEBUG oslo_concurrency.lockutils [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Acquiring lock "refresh_cache-42a9c6f2-08f0-4548-9101-685fc3930c7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.350362] env[65788]: DEBUG oslo_concurrency.lockutils [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Acquired lock "refresh_cache-42a9c6f2-08f0-4548-9101-685fc3930c7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.350362] env[65788]: DEBUG nova.network.neutron [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Refreshing network info cache for port 290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 668.482894] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.483083] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.483253] env[65788]: DEBUG nova.network.neutron [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 668.562142] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.562415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.618744] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b0fd0c-b181-4df5-baa5-b7404b6c320e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.651280] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb84e15-727c-4f0f-8df9-ec2e793d114e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.665806] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661908, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.618157} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.666179] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 42a9c6f2-08f0-4548-9101-685fc3930c7d/42a9c6f2-08f0-4548-9101-685fc3930c7d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 668.667057] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 668.667057] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa65a1f4-57e6-4a96-aea3-15ac9f6247d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.672082] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98557f95-855f-4535-b743-97a086d7594f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.680592] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 668.680592] env[65788]: value = "task-4661909" [ 668.680592] env[65788]: _type = "Task" [ 668.680592] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.691531] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfc0252-631e-4af5-b5cb-eeaa66e3af71 tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "27ce0779-2fa2-44d2-a424-072183c022a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.644s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.692289] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661907, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.694937] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2625ddf-7049-4a4d-9cb2-12197ae85090 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.701852] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.737271] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f17607-3257-4d2a-b54b-fa38ec2db96f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.753500] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296bc06d-0e82-4e7f-8ee1-9c3844f5dbe4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.772913] env[65788]: DEBUG nova.compute.provider_tree [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.780435] env[65788]: DEBUG oslo_concurrency.lockutils [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] Releasing lock "refresh_cache-2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.781214] env[65788]: DEBUG nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Received event network-vif-deleted-d564449f-c908-4d2b-9416-3ea9d3d371fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 668.781214] env[65788]: DEBUG nova.compute.manager [req-9ab645e3-f2b7-434a-908f-2b5cdeef0784 req-1e9881ba-1303-45b4-bf10-09ba98273955 service nova] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Received event network-vif-deleted-528d3b24-3631-4afa-af44-515f60a2df2c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 668.837393] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "86ac6437-511a-4d72-aff0-0325e2d633f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.837629] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.855409] env[65788]: WARNING neutronclient.v2_0.client [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.856355] env[65788]: WARNING openstack [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.856746] env[65788]: WARNING openstack [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.889471] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.987313] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.987728] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.065212] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 669.129665] env[65788]: DEBUG nova.network.neutron [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 669.168276] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 669.169410] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3e678b25-b532-4839-960e-7feb7c8fde55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.184059] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 669.184059] env[65788]: value = "task-4661910" [ 669.184059] env[65788]: _type = "Task" [ 669.184059] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.192314] env[65788]: DEBUG oslo_vmware.api [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661907, 'name': PowerOnVM_Task, 'duration_secs': 2.349986} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.196976] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 669.197180] env[65788]: INFO nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Took 12.50 seconds to spawn the instance on the hypervisor. [ 669.197358] env[65788]: DEBUG nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 669.198890] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb53f40-3334-47f5-8b24-48127fc8424e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.206925] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08208} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.207992] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.210117] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9647ed-8072-4314-8ce6-be42fb6b3222 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.219971] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661910, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.259659] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 42a9c6f2-08f0-4548-9101-685fc3930c7d/42a9c6f2-08f0-4548-9101-685fc3930c7d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.260560] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59dcea0a-681f-4665-adf7-36d166308192 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.280026] env[65788]: WARNING openstack [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.280479] env[65788]: WARNING openstack [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.289788] env[65788]: DEBUG nova.scheduler.client.report [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 669.296462] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.296966] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.307673] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 669.307673] env[65788]: value = "task-4661911" [ 669.307673] env[65788]: _type = "Task" [ 669.307673] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.321034] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661911, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.344811] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 669.418737] env[65788]: WARNING neutronclient.v2_0.client [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.419996] env[65788]: WARNING openstack [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.420137] env[65788]: WARNING openstack [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.452387] env[65788]: WARNING neutronclient.v2_0.client [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 669.453714] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.454362] env[65788]: WARNING openstack [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.606476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.613832] env[65788]: DEBUG nova.network.neutron [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Updated VIF entry in instance network info cache for port 290d5ac2-d3c4-4f0d-b549-523d4ef21f1b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 669.614202] env[65788]: DEBUG nova.network.neutron [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Updating instance_info_cache with network_info: [{"id": "290d5ac2-d3c4-4f0d-b549-523d4ef21f1b", "address": "fa:16:3e:7c:a0:74", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap290d5ac2-d3", "ovs_interfaceid": "290d5ac2-d3c4-4f0d-b549-523d4ef21f1b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.662487] env[65788]: DEBUG nova.network.neutron [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 669.704168] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661910, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.738358] env[65788]: INFO nova.compute.manager [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Took 33.34 seconds to build instance. [ 669.795753] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.829s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.797167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.018s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.797282] env[65788]: DEBUG nova.objects.instance [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 669.822610] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661911, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.874674] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.108104] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.108403] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.117918] env[65788]: DEBUG oslo_concurrency.lockutils [req-a91725db-4c9c-45f0-a3d3-f1b4e09fc6d3 req-3d7f1837-ee68-4166-9253-b9196e575c21 service nova] Releasing lock "refresh_cache-42a9c6f2-08f0-4548-9101-685fc3930c7d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.171061] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.171061] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Instance network_info: |[{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 670.171360] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:3a:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89c8bddf-8224-4440-87da-352af62539fc', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 670.178927] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Creating folder: Project (89909e0b9fd9487982d43248f9022481). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.180018] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9b82086-e815-453b-9eb3-bf5e7335c3ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.194679] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Created folder: Project (89909e0b9fd9487982d43248f9022481) in parent group-v910111. [ 670.194930] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Creating folder: Instances. Parent ref: group-v910182. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 670.199435] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa5a5d44-e321-4ba0-9c6b-b3e3c507d862 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.207555] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661910, 'name': CreateSnapshot_Task, 'duration_secs': 0.598811} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.207895] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 670.208797] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c168e7a-2af5-48ea-aae2-f7e6ccc1f164 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.213578] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Created folder: Instances in parent group-v910182. [ 670.213931] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 670.215303] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 670.215303] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c805542-03ae-4d06-b3be-090b997f6317 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.241416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7f3ce582-838f-42ac-860d-61c8baa3957b tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.858s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.245934] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 670.245934] env[65788]: value = "task-4661915" [ 670.245934] env[65788]: _type = "Task" [ 670.245934] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.257397] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661915, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.303820] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "d61482f5-a0f6-45cb-ab3a-9e4f41ab2dd2" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.304130] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d61482f5-a0f6-45cb-ab3a-9e4f41ab2dd2" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.321011] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661911, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.747581] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 670.748036] env[65788]: DEBUG nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 670.751588] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-90fbd660-28f3-4717-8f47-1e8d46ccec3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.764831] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661915, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.766405] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 670.766405] env[65788]: value = "task-4661916" [ 670.766405] env[65788]: _type = "Task" [ 670.766405] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.775900] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661916, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.811867] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049cc2a9-d749-47c5-8f2c-7d4d468c1f87 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.813447] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d61482f5-a0f6-45cb-ab3a-9e4f41ab2dd2" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.509s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.813954] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 670.816804] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.486s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.817091] env[65788]: DEBUG nova.objects.instance [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lazy-loading 'resources' on Instance uuid 350f3c86-063c-4ed5-895b-fc621b2fa825 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 670.829638] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661911, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.268055] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661915, 'name': CreateVM_Task, 'duration_secs': 0.84474} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.271587] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 671.271820] env[65788]: WARNING neutronclient.v2_0.client [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.272223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.272352] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.272666] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 671.273385] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9da4a223-d994-48d6-8084-b5b5e3d07f3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.279571] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661916, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.281012] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 671.281012] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f10914-e020-d712-4dfa-068fe2dfd7ec" [ 671.281012] env[65788]: _type = "Task" [ 671.281012] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.285187] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.291452] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f10914-e020-d712-4dfa-068fe2dfd7ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.324203] env[65788]: DEBUG nova.compute.utils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 671.331913] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 671.332237] env[65788]: DEBUG nova.network.neutron [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 671.332400] env[65788]: WARNING neutronclient.v2_0.client [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.332662] env[65788]: WARNING neutronclient.v2_0.client [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.333241] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.333581] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.340049] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661911, 'name': ReconfigVM_Task, 'duration_secs': 1.652992} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.341598] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 42a9c6f2-08f0-4548-9101-685fc3930c7d/42a9c6f2-08f0-4548-9101-685fc3930c7d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 671.341825] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ca8be6d-242c-43c0-a531-ae9c46e1a525 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.351481] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 671.351481] env[65788]: value = "task-4661917" [ 671.351481] env[65788]: _type = "Task" [ 671.351481] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.370984] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661917, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.396237] env[65788]: DEBUG nova.policy [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '50bc804533bf4194ad18e280b51d76e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c131f56de9d8479389743350c366e67e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 671.551308] env[65788]: DEBUG nova.compute.manager [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-vif-plugged-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 671.551684] env[65788]: DEBUG oslo_concurrency.lockutils [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Acquiring lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.552099] env[65788]: DEBUG oslo_concurrency.lockutils [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.552260] env[65788]: DEBUG oslo_concurrency.lockutils [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.552503] env[65788]: DEBUG nova.compute.manager [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] No waiting events found dispatching network-vif-plugged-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 671.552714] env[65788]: WARNING nova.compute.manager [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received unexpected event network-vif-plugged-89c8bddf-8224-4440-87da-352af62539fc for instance with vm_state building and task_state spawning. [ 671.552916] env[65788]: DEBUG nova.compute.manager [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-changed-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 671.553143] env[65788]: DEBUG nova.compute.manager [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing instance network info cache due to event network-changed-89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 671.553774] env[65788]: DEBUG oslo_concurrency.lockutils [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Acquiring lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.553774] env[65788]: DEBUG oslo_concurrency.lockutils [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Acquired lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.553774] env[65788]: DEBUG nova.network.neutron [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing network info cache for port 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 671.765677] env[65788]: DEBUG nova.network.neutron [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Successfully created port: 25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 671.781172] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661916, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.788480] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e2fd8f-4f32-4a9f-b5ef-6e5ab56bda67 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.794957] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f10914-e020-d712-4dfa-068fe2dfd7ec, 'name': SearchDatastore_Task, 'duration_secs': 0.034806} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.796858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.796858] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 671.796858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.796858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 671.797218] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 671.797218] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba1a93c7-f8bd-4f95-b1e0-246e23017f78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.801829] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8bbf2d5-38d7-4130-be93-7ea02c7dce84 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.809024] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 671.809313] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 671.853063] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30dcb5e0-d841-48d3-b73f-d12c174429ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.857635] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 671.864141] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6c6576-83c1-4356-9f95-9adac3e36a63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.873972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.873972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.874257] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.874510] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.874728] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.879864] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 671.879864] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52673193-d1d4-cf38-08f7-d9e512561982" [ 671.879864] env[65788]: _type = "Task" [ 671.879864] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.880602] env[65788]: INFO nova.compute.manager [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Terminating instance [ 671.900680] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd72e19d-3e59-4901-834f-70515dae5f01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.907045] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661917, 'name': Rename_Task, 'duration_secs': 0.269442} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.908091] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 671.909083] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1033aa49-653d-44d4-a3c9-94abe205bbc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.917512] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52673193-d1d4-cf38-08f7-d9e512561982, 'name': SearchDatastore_Task, 'duration_secs': 0.016609} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.931514] env[65788]: DEBUG nova.compute.provider_tree [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.935382] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c56e84ab-8c49-440e-b18f-3f0c5291c00b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.941748] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 671.941748] env[65788]: value = "task-4661919" [ 671.941748] env[65788]: _type = "Task" [ 671.941748] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.950613] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 671.950613] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529dcf08-7524-b55a-83dd-7a7d3f72be6b" [ 671.950613] env[65788]: _type = "Task" [ 671.950613] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.961686] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661919, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.971159] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529dcf08-7524-b55a-83dd-7a7d3f72be6b, 'name': SearchDatastore_Task, 'duration_secs': 0.020932} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.971513] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.971834] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1/12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 671.972182] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-523eecb4-6f2f-4a33-80fb-2a77d62886a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.981161] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 671.981161] env[65788]: value = "task-4661920" [ 671.981161] env[65788]: _type = "Task" [ 671.981161] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.994193] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.057631] env[65788]: WARNING neutronclient.v2_0.client [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.058408] env[65788]: WARNING openstack [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.058935] env[65788]: WARNING openstack [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.279451] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661916, 'name': CloneVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.402176] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "refresh_cache-78b545b3-1f84-4ea8-b180-f9039f2ea4b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.402176] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "refresh_cache-78b545b3-1f84-4ea8-b180-f9039f2ea4b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.402176] env[65788]: DEBUG nova.network.neutron [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 672.435823] env[65788]: DEBUG nova.scheduler.client.report [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 672.464528] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661919, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.495798] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661920, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.515021] env[65788]: WARNING openstack [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.515493] env[65788]: WARNING openstack [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.781712] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661916, 'name': CloneVM_Task, 'duration_secs': 1.556589} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.781712] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created linked-clone VM from snapshot [ 672.782013] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca2127a-d574-4a90-a153-b5e955abf596 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.791514] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Uploading image 65b7cb7b-8569-45e3-b201-5d827f11fc08 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 672.826352] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 672.826352] env[65788]: value = "vm-910185" [ 672.826352] env[65788]: _type = "VirtualMachine" [ 672.826352] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 672.826750] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-22d3e361-f3c8-4cfa-b485-26c39bd7af51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.836291] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease: (returnval){ [ 672.836291] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cbaf2d-55b4-2195-ebd9-43e267dccbea" [ 672.836291] env[65788]: _type = "HttpNfcLease" [ 672.836291] env[65788]: } obtained for exporting VM: (result){ [ 672.836291] env[65788]: value = "vm-910185" [ 672.836291] env[65788]: _type = "VirtualMachine" [ 672.836291] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 672.836678] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the lease: (returnval){ [ 672.836678] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cbaf2d-55b4-2195-ebd9-43e267dccbea" [ 672.836678] env[65788]: _type = "HttpNfcLease" [ 672.836678] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 672.845712] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 672.845712] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cbaf2d-55b4-2195-ebd9-43e267dccbea" [ 672.845712] env[65788]: _type = "HttpNfcLease" [ 672.845712] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 672.870924] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 672.899229] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 672.899590] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 672.900241] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 672.900241] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 672.900241] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 672.901170] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 672.901170] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.901170] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 672.901387] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 672.901387] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 672.901573] env[65788]: DEBUG nova.virt.hardware [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 672.902731] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baccb76b-2e51-4980-9c12-89c748e095b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.905972] env[65788]: WARNING neutronclient.v2_0.client [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 672.909040] env[65788]: WARNING openstack [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.909040] env[65788]: WARNING openstack [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.924115] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b74a393-6951-4fbf-9f86-032d5e561078 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.943422] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.126s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 672.947846] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.334s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.949345] env[65788]: INFO nova.compute.claims [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.966025] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661919, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.977140] env[65788]: INFO nova.scheduler.client.report [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Deleted allocations for instance 350f3c86-063c-4ed5-895b-fc621b2fa825 [ 672.994561] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661920, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.661668} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.994866] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1/12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 672.995091] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 672.995372] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9e21ea5-77c9-46a0-8f96-db6055fd8ba7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.004441] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 673.004441] env[65788]: value = "task-4661922" [ 673.004441] env[65788]: _type = "Task" [ 673.004441] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.016726] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661922, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.205185] env[65788]: WARNING neutronclient.v2_0.client [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.205185] env[65788]: WARNING openstack [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.205185] env[65788]: WARNING openstack [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.246116] env[65788]: DEBUG nova.network.neutron [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 673.348608] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 673.348608] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cbaf2d-55b4-2195-ebd9-43e267dccbea" [ 673.348608] env[65788]: _type = "HttpNfcLease" [ 673.348608] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 673.348608] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 673.348608] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cbaf2d-55b4-2195-ebd9-43e267dccbea" [ 673.348608] env[65788]: _type = "HttpNfcLease" [ 673.348608] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 673.348608] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e0d7a3-e43d-4155-8518-c14e39d4b30c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.357917] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5250b045-9ea6-c6b3-49f0-dc8c590a655b/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 673.358113] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5250b045-9ea6-c6b3-49f0-dc8c590a655b/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 673.425074] env[65788]: DEBUG nova.network.neutron [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updated VIF entry in instance network info cache for port 89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 673.425074] env[65788]: DEBUG nova.network.neutron [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.433783] env[65788]: DEBUG nova.network.neutron [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.458733] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661919, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.461918] env[65788]: DEBUG nova.network.neutron [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Successfully updated port: 25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 673.485245] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c80a1eec-890a-4ae4-a604-4facd1459d64 tempest-TenantUsagesTestJSON-1133126431 tempest-TenantUsagesTestJSON-1133126431-project-member] Lock "350f3c86-063c-4ed5-895b-fc621b2fa825" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.598s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.519744] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661922, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.229439} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.519744] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 673.523821] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb722b9a-6891-49b9-87ac-74f73e4a5151 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.547160] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1/12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 673.548022] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80886602-e1c7-40a5-bccc-9d2bd6ad2928 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.574200] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 673.574200] env[65788]: value = "task-4661923" [ 673.574200] env[65788]: _type = "Task" [ 673.574200] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.580615] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7d7a045d-e282-4f00-a922-0d1339db0835 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.590574] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661923, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.928619] env[65788]: DEBUG oslo_concurrency.lockutils [req-6695d249-5ab9-48ab-bf12-741718152043 req-0913cebf-2063-43c4-87c1-6344b0ff7f30 service nova] Releasing lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.936588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "refresh_cache-78b545b3-1f84-4ea8-b180-f9039f2ea4b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.936847] env[65788]: DEBUG nova.compute.manager [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 673.938058] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.938720] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af200ee-228c-4b72-9f42-e3e0f45301f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.949703] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 673.953667] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e3ee62e-888a-47c1-8318-66506830b65d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.964678] env[65788]: DEBUG oslo_vmware.api [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661919, 'name': PowerOnVM_Task, 'duration_secs': 1.842699} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.969777] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 673.970929] env[65788]: INFO nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Took 14.71 seconds to spawn the instance on the hypervisor. [ 673.970929] env[65788]: DEBUG nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 673.971423] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "refresh_cache-d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.971593] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquired lock "refresh_cache-d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 673.971851] env[65788]: DEBUG nova.network.neutron [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 673.973150] env[65788]: DEBUG oslo_vmware.api [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 673.973150] env[65788]: value = "task-4661925" [ 673.973150] env[65788]: _type = "Task" [ 673.973150] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.974539] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535d2337-c074-45fc-b115-f3640c320804 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.996050] env[65788]: DEBUG oslo_vmware.api [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.091980] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661923, 'name': ReconfigVM_Task, 'duration_secs': 0.46239} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.092445] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Reconfigured VM instance instance-00000018 to attach disk [datastore2] 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1/12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 674.093309] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-242d94c4-5e96-4b59-8d50-d840f7c231ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.101811] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 674.101811] env[65788]: value = "task-4661926" [ 674.101811] env[65788]: _type = "Task" [ 674.101811] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.115811] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661926, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.464552] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf893a5-3f2c-4c27-9113-b26511d8cf6e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.478626] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568bcc89-200f-40c4-a1f0-4af9614a03aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.485187] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.486297] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.534848] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443a8eb5-b7dc-4423-a3d9-1efd0d5d17ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.541156] env[65788]: INFO nova.compute.manager [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Took 37.44 seconds to build instance. [ 674.542442] env[65788]: DEBUG oslo_vmware.api [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661925, 'name': PowerOffVM_Task, 'duration_secs': 0.546779} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.543530] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 674.543709] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 674.543994] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9bca227-b1cb-4f09-a466-711277c74432 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.549908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1ff7f4-300d-4272-8b66-7b08796be05a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.566072] env[65788]: DEBUG nova.compute.provider_tree [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.577631] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 674.578212] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 674.578678] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleting the datastore file [datastore2] 78b545b3-1f84-4ea8-b180-f9039f2ea4b1 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 674.580183] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20c2f8ef-9398-44bf-b922-c796b3abef11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.589505] env[65788]: DEBUG oslo_vmware.api [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 674.589505] env[65788]: value = "task-4661928" [ 674.589505] env[65788]: _type = "Task" [ 674.589505] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.601793] env[65788]: DEBUG oslo_vmware.api [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.615616] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661926, 'name': Rename_Task, 'duration_secs': 0.228772} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.615938] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 674.616236] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98f6d238-9bf4-481d-9dc7-bb2f291ff0d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.624905] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 674.624905] env[65788]: value = "task-4661929" [ 674.624905] env[65788]: _type = "Task" [ 674.624905] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.636673] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.645043] env[65788]: DEBUG nova.network.neutron [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 674.683016] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.683302] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.835263] env[65788]: DEBUG nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Received event network-changed-6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 674.835526] env[65788]: DEBUG nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Refreshing instance network info cache due to event network-changed-6209b956-7cf9-4f9d-b25d-a36cca24051a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 674.835798] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Acquiring lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.836194] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Acquired lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.836277] env[65788]: DEBUG nova.network.neutron [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Refreshing network info cache for port 6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 674.886878] env[65788]: WARNING neutronclient.v2_0.client [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 674.887843] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.888618] env[65788]: WARNING openstack [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.044541] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d3c1761-5ff5-4924-bb0d-8763076bf273 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.954s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.072848] env[65788]: DEBUG nova.scheduler.client.report [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 675.108247] env[65788]: DEBUG oslo_vmware.api [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4661928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162628} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.110127] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 675.110622] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 675.110704] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 675.110900] env[65788]: INFO nova.compute.manager [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Took 1.17 seconds to destroy the instance on the hypervisor. [ 675.111923] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 675.111923] env[65788]: DEBUG nova.compute.manager [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 675.111923] env[65788]: DEBUG nova.network.neutron [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 675.112093] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.112686] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.116463] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.139790] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661929, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.190866] env[65788]: DEBUG nova.network.neutron [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Updating instance_info_cache with network_info: [{"id": "25e52706-fea9-4ef1-8374-25b6f71a10d4", "address": "fa:16:3e:2f:4c:1f", "network": {"id": "673d2a91-0df8-4195-95ba-c292d9fe88d1", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-213991661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c131f56de9d8479389743350c366e67e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e52706-fe", "ovs_interfaceid": "25e52706-fea9-4ef1-8374-25b6f71a10d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.201919] env[65788]: DEBUG nova.network.neutron [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 675.202307] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.339649] env[65788]: WARNING neutronclient.v2_0.client [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.340514] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.340916] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.581331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.585729] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 675.590040] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.693s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.591175] env[65788]: DEBUG nova.objects.instance [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 675.625359] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.627315] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.655034] env[65788]: DEBUG oslo_vmware.api [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4661929, 'name': PowerOnVM_Task, 'duration_secs': 0.692755} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.655366] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 675.655564] env[65788]: INFO nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Took 8.64 seconds to spawn the instance on the hypervisor. [ 675.655739] env[65788]: DEBUG nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 675.656686] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13190c63-38fe-47fe-ae5d-100494f55cb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.701533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Releasing lock "refresh_cache-d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 675.701533] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Instance network_info: |[{"id": "25e52706-fea9-4ef1-8374-25b6f71a10d4", "address": "fa:16:3e:2f:4c:1f", "network": {"id": "673d2a91-0df8-4195-95ba-c292d9fe88d1", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-213991661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c131f56de9d8479389743350c366e67e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e52706-fe", "ovs_interfaceid": "25e52706-fea9-4ef1-8374-25b6f71a10d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 675.705555] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2f:4c:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3b7bf7d4-8e0c-4cee-84ba-244e73ef6379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25e52706-fea9-4ef1-8374-25b6f71a10d4', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 675.722816] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Creating folder: Project (c131f56de9d8479389743350c366e67e). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 675.727982] env[65788]: DEBUG nova.network.neutron [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.730175] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.730175] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.730359] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.730752] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.731081] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.733089] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cfde82ab-9dc8-4d36-9f07-e43b0286a4fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.735520] env[65788]: INFO nova.compute.manager [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Terminating instance [ 675.753429] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Created folder: Project (c131f56de9d8479389743350c366e67e) in parent group-v910111. [ 675.754507] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Creating folder: Instances. Parent ref: group-v910186. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 675.756537] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c4d87813-5292-492d-bf44-c18a8f51d7c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.776012] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Created folder: Instances in parent group-v910186. [ 675.777119] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 675.777395] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 675.777694] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-283c2861-6945-4e28-a335-e5ea7d34de16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.800898] env[65788]: WARNING neutronclient.v2_0.client [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 675.800898] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.800898] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.819028] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 675.819028] env[65788]: value = "task-4661932" [ 675.819028] env[65788]: _type = "Task" [ 675.819028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.832663] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661932, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.100442] env[65788]: DEBUG nova.compute.utils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 676.114030] env[65788]: DEBUG nova.network.neutron [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updated VIF entry in instance network info cache for port 6209b956-7cf9-4f9d-b25d-a36cca24051a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 676.114030] env[65788]: DEBUG nova.network.neutron [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updating instance_info_cache with network_info: [{"id": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "address": "fa:16:3e:33:55:aa", "network": {"id": "62f163be-5383-479b-82ff-90e2dec87f17", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1288564708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca3090d0ab424469aa3fa576736bca1b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6209b956-7c", "ovs_interfaceid": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 676.115867] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 676.116209] env[65788]: DEBUG nova.network.neutron [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 676.116772] env[65788]: WARNING neutronclient.v2_0.client [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.118289] env[65788]: WARNING neutronclient.v2_0.client [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.118769] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.119378] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.186414] env[65788]: INFO nova.compute.manager [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Took 37.14 seconds to build instance. [ 676.230566] env[65788]: DEBUG nova.policy [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '219b25263dac44dea162a1e124459475', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6284c3d61bc947869abc3e3aefb14846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 676.238884] env[65788]: INFO nova.compute.manager [-] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Took 1.13 seconds to deallocate network for instance. [ 676.246569] env[65788]: DEBUG nova.compute.manager [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 676.247279] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.248757] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe403eb9-f0f1-43bc-9088-b3b47c26462f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.258545] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 676.258545] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8ae513c-84b7-4bec-9c4a-47ee2873d11c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.268711] env[65788]: DEBUG oslo_vmware.api [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 676.268711] env[65788]: value = "task-4661934" [ 676.268711] env[65788]: _type = "Task" [ 676.268711] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.284977] env[65788]: DEBUG oslo_vmware.api [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.328601] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661932, 'name': CreateVM_Task, 'duration_secs': 0.471908} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.328840] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 676.329463] env[65788]: WARNING neutronclient.v2_0.client [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.329890] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.330108] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.330555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 676.331312] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c09101ca-4b40-4405-a3a2-1d3767754fce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.338458] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 676.338458] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e583ed-2ecd-5547-0724-dc537f85915e" [ 676.338458] env[65788]: _type = "Task" [ 676.338458] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.351217] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e583ed-2ecd-5547-0724-dc537f85915e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.383125] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "27ce0779-2fa2-44d2-a424-072183c022a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.383488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "27ce0779-2fa2-44d2-a424-072183c022a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.383712] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "27ce0779-2fa2-44d2-a424-072183c022a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.383964] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "27ce0779-2fa2-44d2-a424-072183c022a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.384182] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "27ce0779-2fa2-44d2-a424-072183c022a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.388032] env[65788]: INFO nova.compute.manager [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Terminating instance [ 676.603977] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 676.609578] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0840600f-1982-45f4-83d1-5d7f44f1ed54 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.610804] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.834s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.611053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.611273] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 676.611625] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.099s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.611874] env[65788]: DEBUG nova.objects.instance [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'resources' on Instance uuid 141aefe8-1b95-4963-854d-da79ddf143f7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 676.613932] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee494c49-e982-40e3-b5e6-6398cc750ec5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.617469] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Releasing lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.617797] env[65788]: DEBUG nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Received event network-vif-plugged-25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 676.618025] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Acquiring lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.618287] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 676.618486] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.618660] env[65788]: DEBUG nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] No waiting events found dispatching network-vif-plugged-25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 676.618819] env[65788]: WARNING nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Received unexpected event network-vif-plugged-25e52706-fea9-4ef1-8374-25b6f71a10d4 for instance with vm_state building and task_state spawning. [ 676.618993] env[65788]: DEBUG nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Received event network-changed-25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 676.619119] env[65788]: DEBUG nova.compute.manager [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Refreshing instance network info cache due to event network-changed-25e52706-fea9-4ef1-8374-25b6f71a10d4. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 676.619305] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Acquiring lock "refresh_cache-d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.619455] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Acquired lock "refresh_cache-d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.619609] env[65788]: DEBUG nova.network.neutron [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Refreshing network info cache for port 25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 676.628141] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d947b7-169d-4354-951d-47cb3aea0cda {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.647689] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e073896-1739-4888-8835-aaad247b4570 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.657064] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa5564d-dee2-4f69-bec7-4926bffa04e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.691654] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dbc939cd-9429-4f93-9907-9efd3b7c5d8e tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.657s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 676.692214] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178593MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 676.692356] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.735677] env[65788]: DEBUG nova.network.neutron [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Successfully created port: 041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 676.750065] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.783291] env[65788]: DEBUG oslo_vmware.api [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661934, 'name': PowerOffVM_Task, 'duration_secs': 0.30717} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.783291] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 676.783291] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 676.783291] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc78d972-f7ae-438e-a872-47765fc4a49c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.849970] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e583ed-2ecd-5547-0724-dc537f85915e, 'name': SearchDatastore_Task, 'duration_secs': 0.016347} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.850414] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.850654] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 676.850893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.851043] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.851299] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.851676] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fc0cca7-6b93-4386-91e7-ad950f67d1f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.862896] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.863102] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 676.864504] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e889fcb1-5a92-4fe4-9255-a6a70fa6ab24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.868725] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 676.869098] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 676.869407] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Deleting the datastore file [datastore1] 4d8f46cd-1c36-4e43-8110-66e9c991f28d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 676.870377] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2aa2e87-e4fa-4beb-b591-31487dcd800e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.874869] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 676.874869] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520212df-e977-7e2c-fad0-b7f92bc3df9c" [ 676.874869] env[65788]: _type = "Task" [ 676.874869] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.879933] env[65788]: DEBUG oslo_vmware.api [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for the task: (returnval){ [ 676.879933] env[65788]: value = "task-4661936" [ 676.879933] env[65788]: _type = "Task" [ 676.879933] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.888701] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520212df-e977-7e2c-fad0-b7f92bc3df9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.896719] env[65788]: DEBUG nova.compute.manager [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 676.897075] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.897484] env[65788]: DEBUG oslo_vmware.api [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.898628] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af226da7-ab54-486a-99a6-4284d072d753 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.909662] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 676.910076] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba5d72d0-bae2-4097-8865-315b18847ab9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.919710] env[65788]: DEBUG oslo_vmware.api [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 676.919710] env[65788]: value = "task-4661937" [ 676.919710] env[65788]: _type = "Task" [ 676.919710] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.933352] env[65788]: DEBUG oslo_vmware.api [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.124493] env[65788]: WARNING neutronclient.v2_0.client [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.125350] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.126095] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.398384] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520212df-e977-7e2c-fad0-b7f92bc3df9c, 'name': SearchDatastore_Task, 'duration_secs': 0.019161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.398682] env[65788]: DEBUG oslo_vmware.api [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Task: {'id': task-4661936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329844} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.399599] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 677.399749] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 677.399961] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 677.400190] env[65788]: INFO nova.compute.manager [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 677.400498] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 677.401094] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83b24dbc-7b11-43e1-bad2-089ef22c6af6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.403241] env[65788]: DEBUG nova.compute.manager [-] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 677.403241] env[65788]: DEBUG nova.network.neutron [-] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 677.403610] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 677.404074] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.404365] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.418254] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 677.418254] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529c0477-0ecc-f0c9-a5a7-7dc1a1ff1a80" [ 677.418254] env[65788]: _type = "Task" [ 677.418254] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.432064] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529c0477-0ecc-f0c9-a5a7-7dc1a1ff1a80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.436900] env[65788]: DEBUG oslo_vmware.api [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661937, 'name': PowerOffVM_Task, 'duration_secs': 0.387276} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.437290] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 677.437569] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 677.437692] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcddd94f-361b-4779-81ef-a26d2b492e5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.520367] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 677.520591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 677.520753] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Deleting the datastore file [datastore1] 27ce0779-2fa2-44d2-a424-072183c022a6 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 677.521283] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d74ac13d-e805-4990-bc11-829ac5c13805 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.529848] env[65788]: DEBUG oslo_vmware.api [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for the task: (returnval){ [ 677.529848] env[65788]: value = "task-4661939" [ 677.529848] env[65788]: _type = "Task" [ 677.529848] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.543181] env[65788]: DEBUG oslo_vmware.api [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.615176] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 677.621212] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4dc26b-e22e-42d7-8644-f8ca829cf997 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.632819] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3c1343-479e-40b0-aaf5-9e8ab2533d90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.671037] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2346e934-d678-4fa0-a01e-9fe740948e0f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.682260] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac65514-4a55-4f6d-8bb1-3ae2a1b5d357 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.697368] env[65788]: DEBUG nova.compute.provider_tree [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.701509] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 677.701838] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 677.701943] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 677.702138] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 677.702281] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 677.702425] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 677.702630] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.702807] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 677.702990] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 677.703212] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 677.703428] env[65788]: DEBUG nova.virt.hardware [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 677.704347] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3ddda5-4f56-448c-8ab6-ef6cbd3ab7a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.714817] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf08d67-3558-4452-be40-0481b043d2f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.930580] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529c0477-0ecc-f0c9-a5a7-7dc1a1ff1a80, 'name': SearchDatastore_Task, 'duration_secs': 0.018763} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.930862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.931197] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c/d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 677.931457] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60569cee-2c22-4f8b-bd6b-e442c4c0ac02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.939753] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 677.939753] env[65788]: value = "task-4661941" [ 677.939753] env[65788]: _type = "Task" [ 677.939753] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.950585] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661941, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.042788] env[65788]: DEBUG oslo_vmware.api [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Task: {'id': task-4661939, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.311409} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.043448] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 678.043448] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 678.043448] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.043778] env[65788]: INFO nova.compute.manager [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 678.043926] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 678.044328] env[65788]: DEBUG nova.compute.manager [-] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 678.044328] env[65788]: DEBUG nova.network.neutron [-] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 678.044542] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.045073] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.045533] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.138620] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.211028] env[65788]: DEBUG nova.scheduler.client.report [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 678.452313] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661941, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.715502] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.104s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.717999] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.915s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.718283] env[65788]: DEBUG nova.objects.instance [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lazy-loading 'resources' on Instance uuid 15aba949-c04c-4021-add8-2e9d4e4771db {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 678.734984] env[65788]: DEBUG nova.network.neutron [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Successfully updated port: 041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 678.760639] env[65788]: INFO nova.scheduler.client.report [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocations for instance 141aefe8-1b95-4963-854d-da79ddf143f7 [ 678.835136] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 678.952232] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661941, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619417} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.952526] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c/d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 678.952737] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 678.953034] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-296441b8-adc2-4991-93f6-b7a645e020aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.962501] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 678.962501] env[65788]: value = "task-4661942" [ 678.962501] env[65788]: _type = "Task" [ 678.962501] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.973636] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661942, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.237910] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "refresh_cache-06f8fc6f-428c-4e40-8815-f0c78d27fa38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.238583] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquired lock "refresh_cache-06f8fc6f-428c-4e40-8815-f0c78d27fa38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.238785] env[65788]: DEBUG nova.network.neutron [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 679.276707] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0be2118e-3197-47e8-83ca-839dea7088bd tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "141aefe8-1b95-4963-854d-da79ddf143f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.088s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.294503] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.294894] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.477702] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661942, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.211846} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.478137] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.479468] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5deb5534-c33d-412b-a83f-cd40dc93ac3d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.509326] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c/d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 679.513435] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de401e20-2ce2-4453-b010-5cbb646723d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.542609] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 679.542609] env[65788]: value = "task-4661943" [ 679.542609] env[65788]: _type = "Task" [ 679.542609] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.557164] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.732047] env[65788]: DEBUG nova.network.neutron [-] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.743579] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.743970] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.773500] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8b4b52-bfc9-41fa-8a07-ed24db0aea7b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.783732] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca9b88b-92b9-4ccb-b886-27baf71ecb9b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.817924] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937b733d-bf25-446b-8639-bf70834b2f2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.828703] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590015d1-7f45-45ce-8f75-88795845dd52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.845332] env[65788]: DEBUG nova.compute.provider_tree [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.037316] env[65788]: DEBUG nova.network.neutron [-] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.062884] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661943, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.235605] env[65788]: INFO nova.compute.manager [-] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Took 2.83 seconds to deallocate network for instance. [ 680.348965] env[65788]: DEBUG nova.scheduler.client.report [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 680.460350] env[65788]: WARNING neutronclient.v2_0.client [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.460759] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.461145] env[65788]: WARNING openstack [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.541195] env[65788]: DEBUG nova.network.neutron [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 680.545371] env[65788]: INFO nova.compute.manager [-] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Took 2.50 seconds to deallocate network for instance. [ 680.561368] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661943, 'name': ReconfigVM_Task, 'duration_secs': 0.588734} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.561467] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Reconfigured VM instance instance-00000019 to attach disk [datastore2] d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c/d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.562361] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-784efeec-db2b-4bfb-994a-c922566449e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.570161] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 680.570161] env[65788]: value = "task-4661945" [ 680.570161] env[65788]: _type = "Task" [ 680.570161] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.582710] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661945, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.625017] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.625017] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.659652] env[65788]: DEBUG nova.network.neutron [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Updated VIF entry in instance network info cache for port 25e52706-fea9-4ef1-8374-25b6f71a10d4. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 680.659921] env[65788]: DEBUG nova.network.neutron [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Updating instance_info_cache with network_info: [{"id": "25e52706-fea9-4ef1-8374-25b6f71a10d4", "address": "fa:16:3e:2f:4c:1f", "network": {"id": "673d2a91-0df8-4195-95ba-c292d9fe88d1", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-213991661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c131f56de9d8479389743350c366e67e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3b7bf7d4-8e0c-4cee-84ba-244e73ef6379", "external-id": "nsx-vlan-transportzone-423", "segmentation_id": 423, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e52706-fe", "ovs_interfaceid": "25e52706-fea9-4ef1-8374-25b6f71a10d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.712904] env[65788]: WARNING neutronclient.v2_0.client [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 680.713505] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.713843] env[65788]: WARNING openstack [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.747770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.834267] env[65788]: DEBUG nova.network.neutron [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Updating instance_info_cache with network_info: [{"id": "041beec2-3c7a-44fa-9df0-2681c50094b9", "address": "fa:16:3e:68:7c:fd", "network": {"id": "6e87b73c-84f4-49bb-8c0f-309bf7d7a82b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1456122035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6284c3d61bc947869abc3e3aefb14846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3261e15f-7e45-4516-acfd-341bab16e3cf", "external-id": "nsx-vlan-transportzone-783", "segmentation_id": 783, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap041beec2-3c", "ovs_interfaceid": "041beec2-3c7a-44fa-9df0-2681c50094b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.854955] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.137s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.858107] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 19.028s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.886216] env[65788]: INFO nova.scheduler.client.report [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Deleted allocations for instance 15aba949-c04c-4021-add8-2e9d4e4771db [ 681.055542] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.086538] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661945, 'name': Rename_Task, 'duration_secs': 0.23172} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.086838] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 681.087119] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-71db8b5d-27b2-4e0d-9fa8-d4d6411ee5b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.095916] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 681.095916] env[65788]: value = "task-4661946" [ 681.095916] env[65788]: _type = "Task" [ 681.095916] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.107652] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.166419] env[65788]: DEBUG oslo_concurrency.lockutils [req-8e1319c5-1f04-49c2-9a0e-5a15bb0fce3f req-8beda818-4b55-4be6-81e6-5b92da4f9aaf service nova] Releasing lock "refresh_cache-d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.237094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.237094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.237094] env[65788]: DEBUG nova.compute.manager [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 681.238139] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe097153-030d-4580-bc31-482beb41f78a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.248157] env[65788]: DEBUG nova.compute.manager [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 681.249397] env[65788]: DEBUG nova.objects.instance [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lazy-loading 'flavor' on Instance uuid 42a9c6f2-08f0-4548-9101-685fc3930c7d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 681.322426] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.325034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.325034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.325034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.325034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.326018] env[65788]: INFO nova.compute.manager [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Terminating instance [ 681.340570] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Releasing lock "refresh_cache-06f8fc6f-428c-4e40-8815-f0c78d27fa38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.342704] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Instance network_info: |[{"id": "041beec2-3c7a-44fa-9df0-2681c50094b9", "address": "fa:16:3e:68:7c:fd", "network": {"id": "6e87b73c-84f4-49bb-8c0f-309bf7d7a82b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1456122035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6284c3d61bc947869abc3e3aefb14846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3261e15f-7e45-4516-acfd-341bab16e3cf", "external-id": "nsx-vlan-transportzone-783", "segmentation_id": 783, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap041beec2-3c", "ovs_interfaceid": "041beec2-3c7a-44fa-9df0-2681c50094b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 681.343952] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:68:7c:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3261e15f-7e45-4516-acfd-341bab16e3cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '041beec2-3c7a-44fa-9df0-2681c50094b9', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 681.352533] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Creating folder: Project (6284c3d61bc947869abc3e3aefb14846). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 681.353542] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d8f92991-7f2e-43d8-b260-f51727e851b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.366811] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Created folder: Project (6284c3d61bc947869abc3e3aefb14846) in parent group-v910111. [ 681.367292] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Creating folder: Instances. Parent ref: group-v910190. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 681.367368] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e2f4dd0-1b10-416a-a140-08428ce9e4da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.380159] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Created folder: Instances in parent group-v910190. [ 681.380159] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 681.380159] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 681.380159] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31cd4b57-3baa-4d06-94a3-b8acd1bed818 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.408348] env[65788]: DEBUG oslo_concurrency.lockutils [None req-049398b7-cbf7-4926-925f-a7dde4b8d7c7 tempest-ServerDiagnosticsNegativeTest-336669837 tempest-ServerDiagnosticsNegativeTest-336669837-project-member] Lock "15aba949-c04c-4021-add8-2e9d4e4771db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.095s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.414473] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 681.414473] env[65788]: value = "task-4661949" [ 681.414473] env[65788]: _type = "Task" [ 681.414473] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.428990] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661949, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.620090] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661946, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.833132] env[65788]: DEBUG nova.compute.manager [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 681.833748] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.834906] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d542f939-46b1-42e9-be82-ed37128753de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.843077] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06ab693-365a-4798-879d-96c4dfbeb4cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.850153] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 681.851017] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff12adbc-e7f6-4731-bfc8-0880b85c5aa9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.856937] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44842160-8f8f-4b25-8121-f6c1c15dbb41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.861727] env[65788]: DEBUG oslo_vmware.api [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 681.861727] env[65788]: value = "task-4661950" [ 681.861727] env[65788]: _type = "Task" [ 681.861727] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.899172] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02290b22-c66a-4469-be24-166f6a66db8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.909670] env[65788]: DEBUG oslo_vmware.api [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661950, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.923679] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb258fd-b1e4-4dba-931e-552f4e1b8e79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.938349] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661949, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.953551] env[65788]: DEBUG nova.compute.provider_tree [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.121156] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661946, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.166234] env[65788]: DEBUG nova.compute.manager [req-58e9e8fd-92af-4924-b9f5-770dad711733 req-3da5aa9b-4664-4c52-8de3-74f996cf839d service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Received event network-vif-plugged-041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 682.166234] env[65788]: DEBUG oslo_concurrency.lockutils [req-58e9e8fd-92af-4924-b9f5-770dad711733 req-3da5aa9b-4664-4c52-8de3-74f996cf839d service nova] Acquiring lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.166234] env[65788]: DEBUG oslo_concurrency.lockutils [req-58e9e8fd-92af-4924-b9f5-770dad711733 req-3da5aa9b-4664-4c52-8de3-74f996cf839d service nova] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.166234] env[65788]: DEBUG oslo_concurrency.lockutils [req-58e9e8fd-92af-4924-b9f5-770dad711733 req-3da5aa9b-4664-4c52-8de3-74f996cf839d service nova] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.166234] env[65788]: DEBUG nova.compute.manager [req-58e9e8fd-92af-4924-b9f5-770dad711733 req-3da5aa9b-4664-4c52-8de3-74f996cf839d service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] No waiting events found dispatching network-vif-plugged-041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 682.166373] env[65788]: WARNING nova.compute.manager [req-58e9e8fd-92af-4924-b9f5-770dad711733 req-3da5aa9b-4664-4c52-8de3-74f996cf839d service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Received unexpected event network-vif-plugged-041beec2-3c7a-44fa-9df0-2681c50094b9 for instance with vm_state building and task_state spawning. [ 682.219473] env[65788]: DEBUG nova.compute.manager [req-adb16bbf-c549-4b82-b97e-8a2ec8ba2aa5 req-81aabf96-905b-4994-8389-edb4f040f3a3 service nova] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Received event network-vif-deleted-855a1c9a-f74e-4cca-afa1-a8bb99f21513 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 682.265348] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 682.265964] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0831d556-6a29-47ef-8708-5c97959f8e29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.287043] env[65788]: DEBUG oslo_vmware.api [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 682.287043] env[65788]: value = "task-4661952" [ 682.287043] env[65788]: _type = "Task" [ 682.287043] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.298128] env[65788]: DEBUG oslo_vmware.api [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661952, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.378041] env[65788]: DEBUG oslo_vmware.api [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661950, 'name': PowerOffVM_Task, 'duration_secs': 0.279386} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.378041] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 682.378041] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 682.378041] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a88a7bf-8ac0-4943-bbbd-9320a3c059bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.430878] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661949, 'name': CreateVM_Task, 'duration_secs': 0.778905} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.431157] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 682.431744] env[65788]: WARNING neutronclient.v2_0.client [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.432252] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.432401] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 682.432805] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 682.433121] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-247f138a-4638-4b30-867f-cfedb5bf8c86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.442148] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 682.442148] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a59898-4627-c998-9faa-58bf9a0be41e" [ 682.442148] env[65788]: _type = "Task" [ 682.442148] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.454202] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a59898-4627-c998-9faa-58bf9a0be41e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.458305] env[65788]: DEBUG nova.scheduler.client.report [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 682.469383] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 682.469383] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 682.469383] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Deleting the datastore file [datastore2] 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 682.469383] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf29b19b-8dae-4c79-a1b6-d9511c623891 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.483845] env[65788]: DEBUG oslo_vmware.api [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for the task: (returnval){ [ 682.483845] env[65788]: value = "task-4661954" [ 682.483845] env[65788]: _type = "Task" [ 682.483845] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.494797] env[65788]: DEBUG oslo_vmware.api [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661954, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.624848] env[65788]: DEBUG oslo_vmware.api [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661946, 'name': PowerOnVM_Task, 'duration_secs': 1.357405} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.625183] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 682.625375] env[65788]: INFO nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Took 9.75 seconds to spawn the instance on the hypervisor. [ 682.626071] env[65788]: DEBUG nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 682.626446] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3d76ae-707b-4532-885e-325e4588b4b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.673283] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5250b045-9ea6-c6b3-49f0-dc8c590a655b/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 682.674640] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cf3cb3-06c0-4b80-bda6-312f5e58a117 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.683067] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5250b045-9ea6-c6b3-49f0-dc8c590a655b/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 682.683299] env[65788]: ERROR oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5250b045-9ea6-c6b3-49f0-dc8c590a655b/disk-0.vmdk due to incomplete transfer. [ 682.683581] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1333bca8-3331-4f35-9d6c-b86ac2962dd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.693524] env[65788]: DEBUG oslo_vmware.rw_handles [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5250b045-9ea6-c6b3-49f0-dc8c590a655b/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 682.693914] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Uploaded image 65b7cb7b-8569-45e3-b201-5d827f11fc08 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 682.695891] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 682.696326] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-75877f0c-fe60-4e91-9264-8251f0b2ccf7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.705596] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 682.705596] env[65788]: value = "task-4661955" [ 682.705596] env[65788]: _type = "Task" [ 682.705596] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.717589] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661955, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.801548] env[65788]: DEBUG oslo_vmware.api [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661952, 'name': PowerOffVM_Task, 'duration_secs': 0.250696} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.802689] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 682.802893] env[65788]: DEBUG nova.compute.manager [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 682.803908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d93fd1b-163d-470c-afd1-4e062dff153d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.958514] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a59898-4627-c998-9faa-58bf9a0be41e, 'name': SearchDatastore_Task, 'duration_secs': 0.015959} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.958840] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 682.959218] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 682.959468] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.959757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 682.959947] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 682.960299] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b171fb1a-3367-43b5-b246-93e6f24c6374 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.973993] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 682.974472] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 682.975783] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bd58f5a-b4b6-4326-b315-bd2f5450fdd7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.984519] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 682.984519] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b50f06-74f0-1120-f19a-c2aa7a43d504" [ 682.984519] env[65788]: _type = "Task" [ 682.984519] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.001337] env[65788]: DEBUG oslo_vmware.api [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Task: {'id': task-4661954, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.308651} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.006248] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 683.006570] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 683.006828] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.007466] env[65788]: INFO nova.compute.manager [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Took 1.17 seconds to destroy the instance on the hypervisor. [ 683.007551] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 683.008275] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b50f06-74f0-1120-f19a-c2aa7a43d504, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.008599] env[65788]: DEBUG nova.compute.manager [-] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 683.008755] env[65788]: DEBUG nova.network.neutron [-] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 683.009132] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.010122] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.010599] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.105204] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 683.156705] env[65788]: INFO nova.compute.manager [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Took 39.51 seconds to build instance. [ 683.190271] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "1ef08e8f-744d-4939-9433-a4301cd31e77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.190600] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.219480] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661955, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.319027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-91fa71b4-3a91-411f-bdd1-915e6c73cad7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.082s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.473203] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.615s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.476683] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.679s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.478380] env[65788]: INFO nova.compute.claims [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.504939] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b50f06-74f0-1120-f19a-c2aa7a43d504, 'name': SearchDatastore_Task, 'duration_secs': 0.024896} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.506165] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a481adb-3d8b-4af7-b18e-f4fb9886d9be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.514911] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 683.514911] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243d044-3ce3-fe87-ad67-d59e67dc0ecf" [ 683.514911] env[65788]: _type = "Task" [ 683.514911] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.526412] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243d044-3ce3-fe87-ad67-d59e67dc0ecf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.658955] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f7005cfd-b92d-4cbc-a851-e4333d7394ca tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.035s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.693041] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 683.720372] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661955, 'name': Destroy_Task, 'duration_secs': 0.787848} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.720719] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroyed the VM [ 683.721045] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 683.721393] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-567d254f-700b-4150-b060-1eef5e9d0fc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.730413] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 683.730413] env[65788]: value = "task-4661956" [ 683.730413] env[65788]: _type = "Task" [ 683.730413] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.744837] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661956, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.870875] env[65788]: DEBUG nova.network.neutron [-] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 684.028629] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243d044-3ce3-fe87-ad67-d59e67dc0ecf, 'name': SearchDatastore_Task, 'duration_secs': 0.018099} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.028798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 684.030264] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 06f8fc6f-428c-4e40-8815-f0c78d27fa38/06f8fc6f-428c-4e40-8815-f0c78d27fa38.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 684.030264] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0cf37dd6-f8ea-4855-92bc-b49e82928a47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.040284] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 684.040284] env[65788]: value = "task-4661957" [ 684.040284] env[65788]: _type = "Task" [ 684.040284] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.049519] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.062699] env[65788]: INFO nova.scheduler.client.report [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleted allocation for migration 29816798-534f-4aa4-bd9b-815c21495fbd [ 684.142068] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "6aa4b17b-816c-4d84-8f74-a81185f3af65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.142068] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.227216] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.248568] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661956, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.374652] env[65788]: INFO nova.compute.manager [-] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Took 1.37 seconds to deallocate network for instance. [ 684.552039] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496378} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.554992] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 06f8fc6f-428c-4e40-8815-f0c78d27fa38/06f8fc6f-428c-4e40-8815-f0c78d27fa38.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 684.555152] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 684.556235] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-288e251d-3f8e-45f6-a506-4765a3e32c7f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.566896] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 684.566896] env[65788]: value = "task-4661958" [ 684.566896] env[65788]: _type = "Task" [ 684.566896] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.573866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-14f96c7f-fabb-4d69-b0b0-664639713a28 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 30.909s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.581993] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661958, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.643947] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 684.745970] env[65788]: DEBUG oslo_vmware.api [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661956, 'name': RemoveSnapshot_Task, 'duration_secs': 0.742239} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.746190] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 684.746416] env[65788]: INFO nova.compute.manager [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 16.13 seconds to snapshot the instance on the hypervisor. [ 684.881928] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.931090] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d9009c-bcf5-45bd-9c2a-a043a271b66f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.941729] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb605bb4-1861-4b7f-a0e2-c30269814a2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.977710] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5aedc56-9d6f-4172-add2-160ba164e06f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.985044] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "f604c16d-2a86-40d3-9891-5b33309b3047" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.985621] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.990882] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7857520-667b-4f44-8a0a-4c6e5b657587 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.009123] env[65788]: DEBUG nova.compute.provider_tree [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.075814] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661958, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.183709} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.076115] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 685.076975] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0070b67-9442-413d-a21c-7f5221656795 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.104813] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] 06f8fc6f-428c-4e40-8815-f0c78d27fa38/06f8fc6f-428c-4e40-8815-f0c78d27fa38.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 685.105330] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e3bb209-0d4e-463e-ad69-588968c7888a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.127399] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 685.127399] env[65788]: value = "task-4661959" [ 685.127399] env[65788]: _type = "Task" [ 685.127399] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.138242] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.168543] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.306999] env[65788]: DEBUG nova.compute.manager [None req-96397fcc-27c6-4684-8f27-d533bc818aeb tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Found 1 images (rotation: 2) {{(pid=65788) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 685.410956] env[65788]: DEBUG nova.compute.manager [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Received event network-changed-041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 685.410956] env[65788]: DEBUG nova.compute.manager [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Refreshing instance network info cache due to event network-changed-041beec2-3c7a-44fa-9df0-2681c50094b9. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 685.410956] env[65788]: DEBUG oslo_concurrency.lockutils [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Acquiring lock "refresh_cache-06f8fc6f-428c-4e40-8815-f0c78d27fa38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.410956] env[65788]: DEBUG oslo_concurrency.lockutils [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Acquired lock "refresh_cache-06f8fc6f-428c-4e40-8815-f0c78d27fa38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 685.411178] env[65788]: DEBUG nova.network.neutron [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Refreshing network info cache for port 041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 685.430156] env[65788]: DEBUG nova.compute.manager [req-1d3d59e7-63b7-4a8f-997b-f4db2d308ce9 req-757eb59a-1a94-4747-9d55-6be5da8a775d service nova] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Received event network-vif-deleted-b6a744b9-f11d-4c57-8cbd-538c3c7d139c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 685.487491] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 685.513234] env[65788]: DEBUG nova.scheduler.client.report [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 685.639707] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661959, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.914255] env[65788]: WARNING neutronclient.v2_0.client [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 685.915896] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.916710] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.018414] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.019728] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 686.021195] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 686.023603] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.714s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.023803] env[65788]: DEBUG nova.objects.instance [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lazy-loading 'resources' on Instance uuid 54b0945e-c9f5-4726-928e-de1d0ded5aae {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 686.086041] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.086041] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.139684] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661959, 'name': ReconfigVM_Task, 'duration_secs': 0.715191} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.139981] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Reconfigured VM instance instance-0000001a to attach disk [datastore2] 06f8fc6f-428c-4e40-8815-f0c78d27fa38/06f8fc6f-428c-4e40-8815-f0c78d27fa38.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 686.140841] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b453402b-5072-4680-a643-718e2dc95eae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.150020] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 686.150020] env[65788]: value = "task-4661960" [ 686.150020] env[65788]: _type = "Task" [ 686.150020] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.162269] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661960, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.199348] env[65788]: WARNING neutronclient.v2_0.client [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.200046] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.200480] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.384968] env[65788]: DEBUG nova.compute.manager [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 686.386207] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5628590a-ac31-40a4-8109-78df2cacb4d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.390330] env[65788]: DEBUG nova.network.neutron [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Updated VIF entry in instance network info cache for port 041beec2-3c7a-44fa-9df0-2681c50094b9. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 686.390703] env[65788]: DEBUG nova.network.neutron [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Updating instance_info_cache with network_info: [{"id": "041beec2-3c7a-44fa-9df0-2681c50094b9", "address": "fa:16:3e:68:7c:fd", "network": {"id": "6e87b73c-84f4-49bb-8c0f-309bf7d7a82b", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1456122035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6284c3d61bc947869abc3e3aefb14846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3261e15f-7e45-4516-acfd-341bab16e3cf", "external-id": "nsx-vlan-transportzone-783", "segmentation_id": 783, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap041beec2-3c", "ovs_interfaceid": "041beec2-3c7a-44fa-9df0-2681c50094b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 686.528144] env[65788]: DEBUG nova.compute.utils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 686.532805] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 686.533010] env[65788]: DEBUG nova.network.neutron [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 686.533345] env[65788]: WARNING neutronclient.v2_0.client [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.533646] env[65788]: WARNING neutronclient.v2_0.client [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 686.534631] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.534976] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.603238] env[65788]: DEBUG nova.policy [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b0c1ce04a704a3eaf0a0d4dd09f7e08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9674e2a5c86b48db8c865a50331ab846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 686.613654] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquiring lock "54ad3472-0cca-4dca-91f2-8c343f706926" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.613852] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "54ad3472-0cca-4dca-91f2-8c343f706926" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.662811] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661960, 'name': Rename_Task, 'duration_secs': 0.347086} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.663155] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 686.663870] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97d43c82-2191-454d-9783-96c0f8653cea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.670995] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 686.670995] env[65788]: value = "task-4661961" [ 686.670995] env[65788]: _type = "Task" [ 686.670995] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.683339] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661961, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.708381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.708753] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.895760] env[65788]: DEBUG oslo_concurrency.lockutils [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Releasing lock "refresh_cache-06f8fc6f-428c-4e40-8815-f0c78d27fa38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 686.895885] env[65788]: DEBUG nova.compute.manager [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Received event network-changed-6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 686.896386] env[65788]: DEBUG nova.compute.manager [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Refreshing instance network info cache due to event network-changed-6209b956-7cf9-4f9d-b25d-a36cca24051a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 686.896386] env[65788]: DEBUG oslo_concurrency.lockutils [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Acquiring lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.896547] env[65788]: DEBUG oslo_concurrency.lockutils [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Acquired lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.896579] env[65788]: DEBUG nova.network.neutron [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Refreshing network info cache for port 6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 686.903245] env[65788]: INFO nova.compute.manager [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] instance snapshotting [ 686.903470] env[65788]: WARNING nova.compute.manager [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 686.912219] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140d6023-8b06-4753-8986-f3a61147503c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.937378] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca26495f-f0fd-4804-892c-fc1d388b0d46 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.025016] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c51893-07fa-4c61-9ad5-f1792928d1a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.034214] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 687.037904] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed0a4be-ef8f-4d3a-b331-6598b35413a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.071881] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1913bfe-340d-426b-b897-4175e4e04f3e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.080834] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d255495-dc72-4452-a327-1a1798db6933 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.096072] env[65788]: DEBUG nova.compute.provider_tree [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.169963] env[65788]: DEBUG nova.network.neutron [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Successfully created port: 42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 687.185582] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661961, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.385670] env[65788]: DEBUG nova.compute.manager [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 687.389021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35563aec-783c-45dd-8e14-713fabcc2418 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.399241] env[65788]: WARNING neutronclient.v2_0.client [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.400265] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.400969] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.452336] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 687.453694] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4b919f8d-a0a0-4d96-bced-0687b1be8395 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.465765] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 687.465765] env[65788]: value = "task-4661962" [ 687.465765] env[65788]: _type = "Task" [ 687.465765] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.482329] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661962, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.599879] env[65788]: DEBUG nova.scheduler.client.report [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.688315] env[65788]: DEBUG oslo_vmware.api [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661961, 'name': PowerOnVM_Task, 'duration_secs': 0.85773} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.688670] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 687.688865] env[65788]: INFO nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Took 10.07 seconds to spawn the instance on the hypervisor. [ 687.689009] env[65788]: DEBUG nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 687.689972] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535e84a7-f9f3-476e-92bd-3da96fc7d8b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.915513] env[65788]: INFO nova.compute.manager [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] instance snapshotting [ 687.916351] env[65788]: DEBUG nova.objects.instance [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'flavor' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 687.976611] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661962, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.052351] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 688.076654] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 688.076840] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 688.076927] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 688.077130] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 688.077273] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 688.077415] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 688.077617] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.077770] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 688.077928] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 688.078105] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 688.078292] env[65788]: DEBUG nova.virt.hardware [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 688.079269] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4c6af2-62b8-4def-9b8e-6904cc48f46a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.088978] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d703d2c6-5eb6-4640-9c48-906d91a192a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.107085] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.083s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.109699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.047s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 688.109900] env[65788]: DEBUG nova.objects.instance [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 688.133030] env[65788]: INFO nova.scheduler.client.report [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Deleted allocations for instance 54b0945e-c9f5-4726-928e-de1d0ded5aae [ 688.214287] env[65788]: INFO nova.compute.manager [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Took 38.62 seconds to build instance. [ 688.422662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8393880-1fad-4178-b3c4-d879979b2d11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.449692] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b44901-e608-44cb-85cf-bb3a3531d63d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.477965] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661962, 'name': CreateSnapshot_Task, 'duration_secs': 0.688172} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.477965] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 688.478850] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e36d80-6e80-47e0-8f66-d2d95b7d3f9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.642542] env[65788]: DEBUG oslo_concurrency.lockutils [None req-331adeb3-20ca-4651-a29b-dcfaab647fe7 tempest-ServerShowV257Test-1803791157 tempest-ServerShowV257Test-1803791157-project-member] Lock "54b0945e-c9f5-4726-928e-de1d0ded5aae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.147s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.715664] env[65788]: DEBUG oslo_concurrency.lockutils [None req-433d4a32-a981-439c-a4c6-f54dc61e397f tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.137s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.910989] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.911391] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.930890] env[65788]: DEBUG nova.network.neutron [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Successfully updated port: 42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 688.966614] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 688.967451] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bf20b0d3-b010-402c-86b3-a4792acd4a71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.978392] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 688.978392] env[65788]: value = "task-4661963" [ 688.978392] env[65788]: _type = "Task" [ 688.978392] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.990788] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661963, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.002056] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 689.003249] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1bcbc494-c383-4854-a724-2b9648d5db3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.015897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.015897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.015897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.015897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.016334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.017451] env[65788]: INFO nova.compute.manager [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Terminating instance [ 689.027576] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 689.027576] env[65788]: value = "task-4661964" [ 689.027576] env[65788]: _type = "Task" [ 689.027576] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.041887] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661964, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.066020] env[65788]: WARNING neutronclient.v2_0.client [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 689.066020] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.066020] env[65788]: WARNING openstack [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.121223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ca63a0b-b830-496b-8152-cc3aa3ec3729 tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.123045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.092s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.127138] env[65788]: INFO nova.compute.claims [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.215826] env[65788]: DEBUG nova.network.neutron [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updated VIF entry in instance network info cache for port 6209b956-7cf9-4f9d-b25d-a36cca24051a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 689.215826] env[65788]: DEBUG nova.network.neutron [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updating instance_info_cache with network_info: [{"id": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "address": "fa:16:3e:33:55:aa", "network": {"id": "62f163be-5383-479b-82ff-90e2dec87f17", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1288564708-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ca3090d0ab424469aa3fa576736bca1b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6209b956-7c", "ovs_interfaceid": "6209b956-7cf9-4f9d-b25d-a36cca24051a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.220728] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 689.434278] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-bcdf4a61-a2e8-4ca6-92fc-faddf5371290" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.434278] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-bcdf4a61-a2e8-4ca6-92fc-faddf5371290" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.434278] env[65788]: DEBUG nova.network.neutron [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 689.494046] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661963, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.523476] env[65788]: DEBUG nova.compute.manager [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 689.524250] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.524695] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809587f8-82bc-4038-8870-25b3f3269ba3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.540096] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661964, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.543108] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 689.543108] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a17bf472-5b3c-41c7-adf8-9b2745814db2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.552432] env[65788]: DEBUG oslo_vmware.api [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 689.552432] env[65788]: value = "task-4661965" [ 689.552432] env[65788]: _type = "Task" [ 689.552432] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.563198] env[65788]: DEBUG oslo_vmware.api [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.731039] env[65788]: DEBUG oslo_concurrency.lockutils [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] Releasing lock "refresh_cache-957f14fc-23f1-4334-b672-b3a75398f716" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.731039] env[65788]: DEBUG nova.compute.manager [req-12b7d778-94df-4b0f-b867-bb92ee78dccb req-57422e29-c1be-4df2-aa58-2ec04c0a44e0 service nova] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Received event network-vif-deleted-48e98881-cbdc-4bbf-8b72-f5b3eaf5cc0b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 689.751035] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.936696] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.937371] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.990555] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661963, 'name': CreateSnapshot_Task, 'duration_secs': 0.774166} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.991207] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 689.992163] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16ea331-5c6b-4989-b1f4-6cfa4e8c5da4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.042840] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661964, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.065367] env[65788]: DEBUG oslo_vmware.api [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661965, 'name': PowerOffVM_Task, 'duration_secs': 0.282775} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.065670] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 690.065837] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 690.066113] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b30c1682-30d9-4860-ab67-9f6e2c0e016a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.122456] env[65788]: DEBUG nova.network.neutron [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 690.145656] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 690.145868] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 690.146475] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Deleting the datastore file [datastore2] d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.146475] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e7251d7-67d8-42ca-af77-dcbf72cfc79c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.155205] env[65788]: DEBUG oslo_vmware.api [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for the task: (returnval){ [ 690.155205] env[65788]: value = "task-4661967" [ 690.155205] env[65788]: _type = "Task" [ 690.155205] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.168329] env[65788]: DEBUG oslo_vmware.api [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.231185] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.231185] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.358901] env[65788]: WARNING neutronclient.v2_0.client [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.360206] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.360396] env[65788]: WARNING openstack [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.503742] env[65788]: DEBUG nova.network.neutron [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Updating instance_info_cache with network_info: [{"id": "42a7074e-6547-4c88-96e5-693f66c36882", "address": "fa:16:3e:fc:ca:55", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42a7074e-65", "ovs_interfaceid": "42a7074e-6547-4c88-96e5-693f66c36882", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 690.514416] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 690.520258] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-82651f29-b829-458a-8519-3dbb41fd3e21 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.527061] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 690.527061] env[65788]: value = "task-4661968" [ 690.527061] env[65788]: _type = "Task" [ 690.527061] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.550971] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661968, 'name': CloneVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.555014] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661964, 'name': CloneVM_Task, 'duration_secs': 1.185064} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.558458] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Created linked-clone VM from snapshot [ 690.559822] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aadc45-c6a2-4dfa-9e7d-15064acb8d9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.570158] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Uploading image 10256abb-214b-4fd5-97f6-98ef7aba694d {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 690.604670] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 690.604670] env[65788]: value = "vm-910194" [ 690.604670] env[65788]: _type = "VirtualMachine" [ 690.604670] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 690.605895] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d4f75104-d3fd-46e2-bf2b-f9a0b4961134 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.613468] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease: (returnval){ [ 690.613468] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5254080b-1ec7-7887-96e4-ace534c8af60" [ 690.613468] env[65788]: _type = "HttpNfcLease" [ 690.613468] env[65788]: } obtained for exporting VM: (result){ [ 690.613468] env[65788]: value = "vm-910194" [ 690.613468] env[65788]: _type = "VirtualMachine" [ 690.613468] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 690.613817] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the lease: (returnval){ [ 690.613817] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5254080b-1ec7-7887-96e4-ace534c8af60" [ 690.613817] env[65788]: _type = "HttpNfcLease" [ 690.613817] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 690.626602] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 690.626602] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5254080b-1ec7-7887-96e4-ace534c8af60" [ 690.626602] env[65788]: _type = "HttpNfcLease" [ 690.626602] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 690.656500] env[65788]: DEBUG nova.compute.manager [req-799f0803-4697-42a5-af85-b9c16c8c5d5a req-9317d2d2-7812-4170-b78a-f01a1f4be22a service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Received event network-vif-plugged-42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 690.656802] env[65788]: DEBUG oslo_concurrency.lockutils [req-799f0803-4697-42a5-af85-b9c16c8c5d5a req-9317d2d2-7812-4170-b78a-f01a1f4be22a service nova] Acquiring lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.657060] env[65788]: DEBUG oslo_concurrency.lockutils [req-799f0803-4697-42a5-af85-b9c16c8c5d5a req-9317d2d2-7812-4170-b78a-f01a1f4be22a service nova] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.657233] env[65788]: DEBUG oslo_concurrency.lockutils [req-799f0803-4697-42a5-af85-b9c16c8c5d5a req-9317d2d2-7812-4170-b78a-f01a1f4be22a service nova] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.657396] env[65788]: DEBUG nova.compute.manager [req-799f0803-4697-42a5-af85-b9c16c8c5d5a req-9317d2d2-7812-4170-b78a-f01a1f4be22a service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] No waiting events found dispatching network-vif-plugged-42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 690.657555] env[65788]: WARNING nova.compute.manager [req-799f0803-4697-42a5-af85-b9c16c8c5d5a req-9317d2d2-7812-4170-b78a-f01a1f4be22a service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Received unexpected event network-vif-plugged-42a7074e-6547-4c88-96e5-693f66c36882 for instance with vm_state building and task_state spawning. [ 690.668489] env[65788]: DEBUG oslo_vmware.api [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Task: {'id': task-4661967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144045} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.671281] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 690.671521] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 690.671723] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.671983] env[65788]: INFO nova.compute.manager [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Took 1.15 seconds to destroy the instance on the hypervisor. [ 690.672174] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 690.673323] env[65788]: DEBUG nova.compute.manager [-] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 690.673323] env[65788]: DEBUG nova.network.neutron [-] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 690.673323] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.673517] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.673703] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.699513] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30cc92e-022e-4b3e-8c6d-f4bd68011014 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.710508] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70752e9d-1a4e-47d0-b47a-29e07b0c9e08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.750205] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b711bbd-ca8c-41d3-afcc-7216a710935a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.759946] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d2aa0e-2ae7-4b93-8f4c-f4871d87048e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.775067] env[65788]: DEBUG nova.compute.provider_tree [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.844094] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 691.009747] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-bcdf4a61-a2e8-4ca6-92fc-faddf5371290" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.010332] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance network_info: |[{"id": "42a7074e-6547-4c88-96e5-693f66c36882", "address": "fa:16:3e:fc:ca:55", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42a7074e-65", "ovs_interfaceid": "42a7074e-6547-4c88-96e5-693f66c36882", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 691.010732] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:ca:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42a7074e-6547-4c88-96e5-693f66c36882', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 691.018547] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 691.018692] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 691.019616] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd69897c-3584-4939-b921-def72f0d1cce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.044859] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661968, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.046966] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 691.046966] env[65788]: value = "task-4661970" [ 691.046966] env[65788]: _type = "Task" [ 691.046966] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.057807] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661970, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.123807] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 691.123807] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5254080b-1ec7-7887-96e4-ace534c8af60" [ 691.123807] env[65788]: _type = "HttpNfcLease" [ 691.123807] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 691.124494] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 691.124494] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5254080b-1ec7-7887-96e4-ace534c8af60" [ 691.124494] env[65788]: _type = "HttpNfcLease" [ 691.124494] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 691.125903] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc24b124-280a-40e4-a265-1cd9a29c4ab2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.136279] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a69bb7-8a2a-3e5e-80e8-19a08b217aae/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 691.136665] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a69bb7-8a2a-3e5e-80e8-19a08b217aae/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 691.241813] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ee0feda5-cf05-4940-a981-db226b6a036c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.290121] env[65788]: DEBUG nova.scheduler.client.report [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 691.547646] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661968, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.561018] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661970, 'name': CreateVM_Task, 'duration_secs': 0.396847} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.561307] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 691.561889] env[65788]: WARNING neutronclient.v2_0.client [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 691.562497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.562657] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.563155] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 691.563445] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe51465a-0553-4a9e-9119-0660641d618a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.570932] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 691.570932] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522d854d-9f66-03c4-89ed-386912cd259c" [ 691.570932] env[65788]: _type = "Task" [ 691.570932] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.582533] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522d854d-9f66-03c4-89ed-386912cd259c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.777083] env[65788]: DEBUG nova.network.neutron [-] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 691.799027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.674s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.799027] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 691.804732] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.915s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 691.808508] env[65788]: INFO nova.compute.claims [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.059158] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661968, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.091684] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522d854d-9f66-03c4-89ed-386912cd259c, 'name': SearchDatastore_Task, 'duration_secs': 0.011423} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.092217] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.095077] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 692.095077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.095077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.095077] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 692.095077] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-374f7be2-b231-43be-a57e-5f62ebc207ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.105803] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 692.106014] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 692.106810] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-204232ee-c69d-4beb-8933-b85db3b87ec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.115862] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 692.115862] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ca2426-cdca-e4a0-e728-a8878404130d" [ 692.115862] env[65788]: _type = "Task" [ 692.115862] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.129617] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ca2426-cdca-e4a0-e728-a8878404130d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.280956] env[65788]: INFO nova.compute.manager [-] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Took 1.61 seconds to deallocate network for instance. [ 692.320280] env[65788]: DEBUG nova.compute.utils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 692.324631] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 692.324631] env[65788]: DEBUG nova.network.neutron [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 692.324631] env[65788]: WARNING neutronclient.v2_0.client [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.324631] env[65788]: WARNING neutronclient.v2_0.client [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 692.325060] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.325436] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.388790] env[65788]: DEBUG nova.policy [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c4a00121e214dc9a7b0866c8deee18e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffbe44d63c1d432e97849f15615329e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 692.559409] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661968, 'name': CloneVM_Task, 'duration_secs': 1.725349} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.559581] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created linked-clone VM from snapshot [ 692.560352] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec47a07-c967-4c9e-ac20-ff7ad90bc3c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.570978] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Uploading image d238e3ba-e173-4cb3-b3f0-97cf0daf02f4 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 692.601213] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 692.601213] env[65788]: value = "vm-910196" [ 692.601213] env[65788]: _type = "VirtualMachine" [ 692.601213] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 692.601837] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5ed256d6-7ad7-4702-93aa-5c345d8f98d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.611529] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease: (returnval){ [ 692.611529] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c41b0-9acc-6516-4335-f9597e0bf7be" [ 692.611529] env[65788]: _type = "HttpNfcLease" [ 692.611529] env[65788]: } obtained for exporting VM: (result){ [ 692.611529] env[65788]: value = "vm-910196" [ 692.611529] env[65788]: _type = "VirtualMachine" [ 692.611529] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 692.611529] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the lease: (returnval){ [ 692.611529] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c41b0-9acc-6516-4335-f9597e0bf7be" [ 692.611529] env[65788]: _type = "HttpNfcLease" [ 692.611529] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 692.625127] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 692.625127] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c41b0-9acc-6516-4335-f9597e0bf7be" [ 692.625127] env[65788]: _type = "HttpNfcLease" [ 692.625127] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 692.633407] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ca2426-cdca-e4a0-e728-a8878404130d, 'name': SearchDatastore_Task, 'duration_secs': 0.012178} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.634471] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd8803b3-dc91-47c0-afb4-55cbb45e1c74 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.645660] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 692.645660] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a65472-9857-da70-039b-d3b053693d70" [ 692.645660] env[65788]: _type = "Task" [ 692.645660] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.660158] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a65472-9857-da70-039b-d3b053693d70, 'name': SearchDatastore_Task, 'duration_secs': 0.013805} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.660158] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.660317] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 692.660642] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c58f6a8-cc65-4e26-933a-640cf43e397b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.670236] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 692.670236] env[65788]: value = "task-4661972" [ 692.670236] env[65788]: _type = "Task" [ 692.670236] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.682736] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.791932] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.828251] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "957f14fc-23f1-4334-b672-b3a75398f716" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.828635] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "957f14fc-23f1-4334-b672-b3a75398f716" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.828920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "957f14fc-23f1-4334-b672-b3a75398f716-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.829151] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "957f14fc-23f1-4334-b672-b3a75398f716-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 692.829396] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "957f14fc-23f1-4334-b672-b3a75398f716-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 692.832642] env[65788]: INFO nova.compute.manager [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Terminating instance [ 692.834704] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 692.930151] env[65788]: DEBUG nova.compute.manager [req-b698343b-f348-446a-bab0-a83c34162dfb req-197ee0e5-cd1d-415d-bb7d-c6e50c312a40 service nova] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Received event network-vif-deleted-25e52706-fea9-4ef1-8374-25b6f71a10d4 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 693.004872] env[65788]: DEBUG nova.network.neutron [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Successfully created port: 02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 693.127247] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 693.127247] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c41b0-9acc-6516-4335-f9597e0bf7be" [ 693.127247] env[65788]: _type = "HttpNfcLease" [ 693.127247] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 693.127525] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 693.127525] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c41b0-9acc-6516-4335-f9597e0bf7be" [ 693.127525] env[65788]: _type = "HttpNfcLease" [ 693.127525] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 693.129231] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084bbabd-53cf-4d6b-8c4d-816fd57b403d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.144457] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290a0f3-3f9f-bbe6-8628-b4e2dd1947df/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 693.144831] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290a0f3-3f9f-bbe6-8628-b4e2dd1947df/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 693.243637] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661972, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.283949] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e5fee90f-6ba3-478d-9533-ee8926fe3824 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.344587] env[65788]: DEBUG nova.compute.manager [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 693.344851] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.345787] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11701735-2d0b-4af7-841f-a789fed17510 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.355390] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 693.358687] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e22c19af-c28a-441a-bd1b-ce2997b42601 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.373103] env[65788]: DEBUG oslo_vmware.api [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 693.373103] env[65788]: value = "task-4661976" [ 693.373103] env[65788]: _type = "Task" [ 693.373103] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.393045] env[65788]: DEBUG oslo_vmware.api [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.494380] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9a7be4-8d00-4a7f-bc35-d806eb48a6b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.504730] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c371790d-de38-403f-813b-e34dc047c0c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.544647] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9089a1-a1f5-4a1c-a1ca-75433339e349 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.556025] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a38cd0-a9e4-4976-994f-4524e426deb4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.574886] env[65788]: DEBUG nova.compute.provider_tree [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.735640] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58048} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.735640] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.736245] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.736600] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c08386b-59b7-4d0d-a186-10b194cf1ed3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.745599] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 693.745599] env[65788]: value = "task-4661977" [ 693.745599] env[65788]: _type = "Task" [ 693.745599] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.757600] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.779829] env[65788]: DEBUG nova.compute.manager [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Received event network-changed-42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 693.779997] env[65788]: DEBUG nova.compute.manager [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Refreshing instance network info cache due to event network-changed-42a7074e-6547-4c88-96e5-693f66c36882. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 693.780353] env[65788]: DEBUG oslo_concurrency.lockutils [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Acquiring lock "refresh_cache-bcdf4a61-a2e8-4ca6-92fc-faddf5371290" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.780532] env[65788]: DEBUG oslo_concurrency.lockutils [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Acquired lock "refresh_cache-bcdf4a61-a2e8-4ca6-92fc-faddf5371290" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.780682] env[65788]: DEBUG nova.network.neutron [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Refreshing network info cache for port 42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 693.852031] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 693.890462] env[65788]: DEBUG oslo_vmware.api [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661976, 'name': PowerOffVM_Task, 'duration_secs': 0.271563} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.891404] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 693.892069] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 693.892561] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22b12a48-ee72-4e66-9fb7-df66761b5beb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.972158] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 693.972390] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 693.973172] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Deleting the datastore file [datastore1] 957f14fc-23f1-4334-b672-b3a75398f716 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 693.973172] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2820f146-aa21-4bab-8725-939caef22f9b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.981701] env[65788]: DEBUG oslo_vmware.api [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for the task: (returnval){ [ 693.981701] env[65788]: value = "task-4661979" [ 693.981701] env[65788]: _type = "Task" [ 693.981701] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.991354] env[65788]: DEBUG oslo_vmware.api [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661979, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.079298] env[65788]: DEBUG nova.scheduler.client.report [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 694.260146] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084578} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.260935] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 694.262437] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6283909c-ac01-46d1-a1cb-c551a77c62fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.292646] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 694.293623] env[65788]: WARNING neutronclient.v2_0.client [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.294294] env[65788]: WARNING openstack [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.295070] env[65788]: WARNING openstack [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.303415] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47316f1f-a687-421b-8f85-5ba5dc31c0da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.328302] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 694.328302] env[65788]: value = "task-4661980" [ 694.328302] env[65788]: _type = "Task" [ 694.328302] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.343786] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661980, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.480994] env[65788]: WARNING openstack [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.483256] env[65788]: WARNING openstack [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.503365] env[65788]: DEBUG oslo_vmware.api [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Task: {'id': task-4661979, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29169} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.503814] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 694.504612] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 694.504612] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.508028] env[65788]: INFO nova.compute.manager [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Took 1.16 seconds to destroy the instance on the hypervisor. [ 694.508028] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 694.508028] env[65788]: DEBUG nova.compute.manager [-] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 694.508028] env[65788]: DEBUG nova.network.neutron [-] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 694.508028] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.508028] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.508532] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.522263] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "72111766-217d-469f-ad92-ca0ce0e67090" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.522752] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "72111766-217d-469f-ad92-ca0ce0e67090" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.585305] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.781s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.586544] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 694.590610] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.984s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.593741] env[65788]: INFO nova.compute.claims [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.599019] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.634632] env[65788]: WARNING neutronclient.v2_0.client [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 694.637517] env[65788]: WARNING openstack [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.637934] env[65788]: WARNING openstack [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.835821] env[65788]: DEBUG nova.network.neutron [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Successfully updated port: 02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 694.849051] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.973992] env[65788]: DEBUG nova.network.neutron [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Updated VIF entry in instance network info cache for port 42a7074e-6547-4c88-96e5-693f66c36882. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 694.974508] env[65788]: DEBUG nova.network.neutron [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Updating instance_info_cache with network_info: [{"id": "42a7074e-6547-4c88-96e5-693f66c36882", "address": "fa:16:3e:fc:ca:55", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42a7074e-65", "ovs_interfaceid": "42a7074e-6547-4c88-96e5-693f66c36882", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.101142] env[65788]: DEBUG nova.compute.utils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 695.104347] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 695.104784] env[65788]: DEBUG nova.network.neutron [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 695.106403] env[65788]: WARNING neutronclient.v2_0.client [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 695.107350] env[65788]: WARNING neutronclient.v2_0.client [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 695.108050] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.108505] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.161164] env[65788]: DEBUG nova.policy [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95f16b049d1f47ab9ecad176bbde147e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '297851f2d1af4507a43948f51d2f088b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 695.342050] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.342050] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.342050] env[65788]: DEBUG nova.network.neutron [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 695.345949] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.394868] env[65788]: DEBUG nova.network.neutron [-] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.477870] env[65788]: DEBUG oslo_concurrency.lockutils [req-c35ef9a3-356a-4887-850e-b5afd68df8ac req-f6b43d1a-dce8-42b9-8f3e-c3154eb44f5f service nova] Releasing lock "refresh_cache-bcdf4a61-a2e8-4ca6-92fc-faddf5371290" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.576286] env[65788]: DEBUG nova.network.neutron [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Successfully created port: f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 695.618910] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 695.846454] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661980, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.847604] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.847726] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.899504] env[65788]: INFO nova.compute.manager [-] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Took 1.39 seconds to deallocate network for instance. [ 695.991631] env[65788]: DEBUG nova.network.neutron [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 696.040246] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.040710] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.118923] env[65788]: WARNING neutronclient.v2_0.client [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 696.119592] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.119936] env[65788]: WARNING openstack [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.181206] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78e8407-5596-4216-9b38-af2050a673a1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.191619] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4043ef-15bc-4ca1-b0c5-f56584480787 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.230142] env[65788]: DEBUG nova.network.neutron [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Updating instance_info_cache with network_info: [{"id": "02852d45-877c-40ee-ade8-47cadad62cf6", "address": "fa:16:3e:cb:04:51", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02852d45-87", "ovs_interfaceid": "02852d45-877c-40ee-ade8-47cadad62cf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.232106] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c6fd86-cc05-4186-b710-8eaf3fc8fe8b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.241841] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061a3c89-44cb-4dd3-90c5-0a8b378b3a24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.259050] env[65788]: DEBUG nova.compute.provider_tree [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.344751] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661980, 'name': ReconfigVM_Task, 'duration_secs': 1.594994} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.344751] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Reconfigured VM instance instance-0000001b to attach disk [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 696.345409] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8159f4e3-37d1-4a9c-b36e-cf63843b24a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.354877] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 696.354877] env[65788]: value = "task-4661981" [ 696.354877] env[65788]: _type = "Task" [ 696.354877] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.366631] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661981, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.407020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.641334] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 696.736978] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.736978] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance network_info: |[{"id": "02852d45-877c-40ee-ade8-47cadad62cf6", "address": "fa:16:3e:cb:04:51", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02852d45-87", "ovs_interfaceid": "02852d45-877c-40ee-ade8-47cadad62cf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 696.762874] env[65788]: DEBUG nova.scheduler.client.report [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 696.868093] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661981, 'name': Rename_Task, 'duration_secs': 0.175161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.868404] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 696.868743] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2bca0c73-9fc0-49e6-80f3-3cddb5399b89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.877859] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 696.877859] env[65788]: value = "task-4661982" [ 696.877859] env[65788]: _type = "Task" [ 696.877859] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.887880] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.252588] env[65788]: DEBUG nova.network.neutron [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Successfully updated port: f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 697.270715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.680s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 697.271528] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 697.275894] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.401s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.278418] env[65788]: INFO nova.compute.claims [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.392584] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661982, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.755861] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.756101] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquired lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.756359] env[65788]: DEBUG nova.network.neutron [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 697.784879] env[65788]: DEBUG nova.compute.utils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 697.790307] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 697.790645] env[65788]: DEBUG nova.network.neutron [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 697.791029] env[65788]: WARNING neutronclient.v2_0.client [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.791529] env[65788]: WARNING neutronclient.v2_0.client [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 697.792466] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.792466] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.860739] env[65788]: DEBUG nova.policy [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '132239fc977a49cd93f59d6706f6794a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c963c936cf20402d9a65f5e61a11020a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 697.891183] env[65788]: DEBUG oslo_vmware.api [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4661982, 'name': PowerOnVM_Task, 'duration_secs': 0.831754} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.891556] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 697.891910] env[65788]: INFO nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Took 9.84 seconds to spawn the instance on the hypervisor. [ 697.892177] env[65788]: DEBUG nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 697.893123] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9993df7-1235-4fe9-b7be-df7102ec2753 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.230866] env[65788]: DEBUG nova.network.neutron [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Successfully created port: fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 698.261559] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.262557] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.291080] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 698.422205] env[65788]: INFO nova.compute.manager [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Took 34.67 seconds to build instance. [ 698.599657] env[65788]: DEBUG nova.network.neutron [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 698.621524] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.621982] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.710261] env[65788]: WARNING neutronclient.v2_0.client [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 698.712023] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.712023] env[65788]: WARNING openstack [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.826965] env[65788]: DEBUG nova.network.neutron [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updating instance_info_cache with network_info: [{"id": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "address": "fa:16:3e:4d:7b:e2", "network": {"id": "833e9a86-34c6-4e5f-b367-630030497015", "bridge": "br-int", "label": "tempest-ServersTestJSON-1044181598-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "297851f2d1af4507a43948f51d2f088b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24d1ba6-c0", "ovs_interfaceid": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 698.849474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b25fbb-f16c-4421-ba44-87392811f3c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.860575] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7256d786-b9dc-46a7-bcdf-03a396827eef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.899848] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef9a6a2-4e10-4a7d-bba3-8124db7767a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.909375] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e3871b-cba6-4b29-ab2b-2787647d6be3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.928475] env[65788]: DEBUG oslo_concurrency.lockutils [None req-60560c6c-398b-4ddb-a4b5-28e588aca20b tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.983s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.929118] env[65788]: DEBUG nova.compute.provider_tree [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.302033] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 699.330827] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Releasing lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.331290] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Instance network_info: |[{"id": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "address": "fa:16:3e:4d:7b:e2", "network": {"id": "833e9a86-34c6-4e5f-b367-630030497015", "bridge": "br-int", "label": "tempest-ServersTestJSON-1044181598-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "297851f2d1af4507a43948f51d2f088b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24d1ba6-c0", "ovs_interfaceid": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 699.433051] env[65788]: DEBUG nova.scheduler.client.report [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 699.437797] env[65788]: DEBUG nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 699.588400] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 699.588710] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 699.588889] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 699.589545] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 699.589751] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 699.590018] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 699.591054] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.591054] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 699.591054] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 699.591054] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 699.591348] env[65788]: DEBUG nova.virt.hardware [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 699.594208] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e80ce12-3d20-4d37-a175-314190b8185c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.608186] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e831c4bb-ca50-4f60-bf53-8f4fef9b8311 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.627977] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:04:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02852d45-877c-40ee-ade8-47cadad62cf6', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.635392] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 699.635533] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 699.636056] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1175c23-b281-4971-82cf-7847f36cb3d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.658278] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.658278] env[65788]: value = "task-4661983" [ 699.658278] env[65788]: _type = "Task" [ 699.658278] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.671959] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661983, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.698557] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 699.699383] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 699.699383] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 699.699383] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 699.699636] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 699.699636] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 699.700295] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.700295] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 699.700295] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 699.700426] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 699.700544] env[65788]: DEBUG nova.virt.hardware [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 699.702905] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a042a84-2f0c-4a4d-a12f-8ecce0313cb3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.717561] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3caec0e-b576-45af-8181-9706aa97f57a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.734642] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:7b:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32463b6d-4569-4755-8a29-873a028690a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.742416] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Creating folder: Project (297851f2d1af4507a43948f51d2f088b). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.747021] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 699.747021] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 699.747021] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 699.747328] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 699.747328] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 699.747328] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 699.747328] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.747328] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 699.747488] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 699.747488] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 699.747488] env[65788]: DEBUG nova.virt.hardware [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 699.747488] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2f49cda-0b62-4ec6-9f9a-9e11b45ad84f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.751077] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9afe1fa-7915-49f1-9e86-bd3dc141aee7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.760330] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b4ec03-62b1-4f1d-af12-693de5a172eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.765625] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Created folder: Project (297851f2d1af4507a43948f51d2f088b) in parent group-v910111. [ 699.765819] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Creating folder: Instances. Parent ref: group-v910202. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.766490] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c575bacd-a716-480c-9848-691527e8e185 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.778965] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Created folder: Instances in parent group-v910202. [ 699.779249] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 699.779457] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 699.779702] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc4bfd8c-4859-4731-8506-32165e98cb4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.802433] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.802433] env[65788]: value = "task-4661986" [ 699.802433] env[65788]: _type = "Task" [ 699.802433] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.812327] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661986, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.924489] env[65788]: DEBUG nova.network.neutron [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Successfully updated port: fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 699.927025] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a69bb7-8a2a-3e5e-80e8-19a08b217aae/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 699.928177] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e615b48e-c174-462a-891d-cf7a322450ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.936611] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a69bb7-8a2a-3e5e-80e8-19a08b217aae/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 699.936611] env[65788]: ERROR oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a69bb7-8a2a-3e5e-80e8-19a08b217aae/disk-0.vmdk due to incomplete transfer. [ 699.936981] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-07c562fa-ce42-4887-94c4-5a5adc8fbd59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.942900] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.666s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.943376] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 699.950602] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.665s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.953708] env[65788]: INFO nova.compute.claims [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.957101] env[65788]: DEBUG oslo_vmware.rw_handles [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a69bb7-8a2a-3e5e-80e8-19a08b217aae/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 699.957101] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Uploaded image 10256abb-214b-4fd5-97f6-98ef7aba694d to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 699.959183] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 699.961673] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-544a7d17-7a14-4d14-b525-8a1b9c4bee9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.971193] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 699.971193] env[65788]: value = "task-4661987" [ 699.971193] env[65788]: _type = "Task" [ 699.971193] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.984903] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661987, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.986290] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.139889] env[65788]: DEBUG nova.compute.manager [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Received event network-vif-plugged-02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 700.140165] env[65788]: DEBUG oslo_concurrency.lockutils [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Acquiring lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.140743] env[65788]: DEBUG oslo_concurrency.lockutils [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.140963] env[65788]: DEBUG oslo_concurrency.lockutils [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 700.141162] env[65788]: DEBUG nova.compute.manager [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] No waiting events found dispatching network-vif-plugged-02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 700.141349] env[65788]: WARNING nova.compute.manager [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Received unexpected event network-vif-plugged-02852d45-877c-40ee-ade8-47cadad62cf6 for instance with vm_state building and task_state spawning. [ 700.141501] env[65788]: DEBUG nova.compute.manager [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Received event network-changed-02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 700.141710] env[65788]: DEBUG nova.compute.manager [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Refreshing instance network info cache due to event network-changed-02852d45-877c-40ee-ade8-47cadad62cf6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 700.141910] env[65788]: DEBUG oslo_concurrency.lockutils [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Acquiring lock "refresh_cache-4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.142057] env[65788]: DEBUG oslo_concurrency.lockutils [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Acquired lock "refresh_cache-4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.142231] env[65788]: DEBUG nova.network.neutron [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Refreshing network info cache for port 02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 700.172076] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661983, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.318578] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661986, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.433449] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.436831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquired lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.436831] env[65788]: DEBUG nova.network.neutron [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 700.451871] env[65788]: DEBUG nova.compute.utils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 700.454762] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 700.454762] env[65788]: DEBUG nova.network.neutron [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 700.455246] env[65788]: WARNING neutronclient.v2_0.client [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.455665] env[65788]: WARNING neutronclient.v2_0.client [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.456522] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.457216] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.487732] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661987, 'name': Destroy_Task, 'duration_secs': 0.44566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.492282] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Destroyed the VM [ 700.492282] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 700.492282] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-270cc440-ccee-4bee-9ce6-6caa558e3567 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.501356] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 700.501356] env[65788]: value = "task-4661988" [ 700.501356] env[65788]: _type = "Task" [ 700.501356] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.508540] env[65788]: DEBUG nova.compute.manager [req-5a6e32e0-a762-4202-8762-3ce93ca6d135 req-5efe755a-460e-4edc-9edc-5e5c25f9b2fd service nova] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Received event network-vif-deleted-6209b956-7cf9-4f9d-b25d-a36cca24051a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 700.517754] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661988, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.556717] env[65788]: DEBUG nova.policy [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c028d9c053054a2f916507b237d37e60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06b3dfec7c0144928b4a59dbd27569b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 700.645668] env[65788]: WARNING neutronclient.v2_0.client [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.646467] env[65788]: WARNING openstack [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.646779] env[65788]: WARNING openstack [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.672267] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661983, 'name': CreateVM_Task, 'duration_secs': 0.535161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.672470] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 700.673173] env[65788]: WARNING neutronclient.v2_0.client [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.673607] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.673749] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.674097] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 700.674395] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd6882d7-ea9b-4139-ad0b-71fc955f391f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.680235] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 700.680235] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a596c5-3c54-d14e-18ac-bad6fcff1216" [ 700.680235] env[65788]: _type = "Task" [ 700.680235] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.692552] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a596c5-3c54-d14e-18ac-bad6fcff1216, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.779134] env[65788]: WARNING openstack [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.779727] env[65788]: WARNING openstack [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.814716] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661986, 'name': CreateVM_Task, 'duration_secs': 0.606627} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.814887] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 700.815617] env[65788]: WARNING neutronclient.v2_0.client [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.815782] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.896328] env[65788]: DEBUG nova.network.neutron [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Successfully created port: 83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 700.941898] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.942380] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.972120] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 700.981056] env[65788]: WARNING neutronclient.v2_0.client [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 700.981056] env[65788]: WARNING openstack [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.982163] env[65788]: WARNING openstack [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.008137] env[65788]: DEBUG nova.network.neutron [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 701.014129] env[65788]: DEBUG oslo_vmware.api [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4661988, 'name': RemoveSnapshot_Task, 'duration_secs': 0.403074} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.014559] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 701.014886] env[65788]: INFO nova.compute.manager [None req-88f6c271-8c45-48bd-86e2-ac61046db50a tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Took 14.10 seconds to snapshot the instance on the hypervisor. [ 701.127345] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.127545] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.174445] env[65788]: DEBUG nova.network.neutron [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Updated VIF entry in instance network info cache for port 02852d45-877c-40ee-ade8-47cadad62cf6. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 701.174714] env[65788]: DEBUG nova.network.neutron [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Updating instance_info_cache with network_info: [{"id": "02852d45-877c-40ee-ade8-47cadad62cf6", "address": "fa:16:3e:cb:04:51", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02852d45-87", "ovs_interfaceid": "02852d45-877c-40ee-ade8-47cadad62cf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 701.196277] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a596c5-3c54-d14e-18ac-bad6fcff1216, 'name': SearchDatastore_Task, 'duration_secs': 0.01747} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.196608] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.197065] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.197409] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.197632] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.198032] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.198403] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.198700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 701.199298] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-538b9b8d-ae2d-4e00-8482-347e24042533 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.201753] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-109a1490-c2a1-44dc-8508-f51f17d1d1f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.212577] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 701.212577] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e639f-6614-2bf3-4df0-9044e9f1818a" [ 701.212577] env[65788]: _type = "Task" [ 701.212577] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.218053] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.218053] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 701.219919] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-749258a6-4662-49cb-96cf-10c7d9e69014 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.226339] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e639f-6614-2bf3-4df0-9044e9f1818a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.231240] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 701.231240] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52984977-be57-5f5e-b142-0c86b0e51e59" [ 701.231240] env[65788]: _type = "Task" [ 701.231240] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.254864] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52984977-be57-5f5e-b142-0c86b0e51e59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.279020] env[65788]: WARNING neutronclient.v2_0.client [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 701.280392] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.280392] env[65788]: WARNING openstack [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.457342] env[65788]: DEBUG nova.network.neutron [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Updating instance_info_cache with network_info: [{"id": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "address": "fa:16:3e:05:d7:71", "network": {"id": "ef52470a-8722-4701-88f2-cb7ecd27ef6b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-878922047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c963c936cf20402d9a65f5e61a11020a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcafb23d-cf", "ovs_interfaceid": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 701.549027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1aa00f-3dee-4cf7-90ad-614bbaa3077b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.558990] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3d935a-8c90-42aa-a39d-892bc02050a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.599284] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f9e57d-72af-462c-bc51-54bb9cc47093 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.610572] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f6a3f1-1478-402f-a8ee-bb6638d8ae17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.628373] env[65788]: DEBUG nova.compute.provider_tree [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.678606] env[65788]: DEBUG oslo_concurrency.lockutils [req-b2c81c12-686b-467a-9003-35c9d2d2ed07 req-32bda4fa-3088-4654-ab29-8a9c1aa9afe9 service nova] Releasing lock "refresh_cache-4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.726998] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e639f-6614-2bf3-4df0-9044e9f1818a, 'name': SearchDatastore_Task, 'duration_secs': 0.062272} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.727799] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.728065] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.728288] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.749166] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.749604] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.749823] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.750014] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.750186] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.752750] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52984977-be57-5f5e-b142-0c86b0e51e59, 'name': SearchDatastore_Task, 'duration_secs': 0.04089} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.752963] env[65788]: INFO nova.compute.manager [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Terminating instance [ 701.756546] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290a0f3-3f9f-bbe6-8628-b4e2dd1947df/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 701.758236] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfd83cb-1fff-41a7-a7ec-95ea77cf96fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.762847] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04fdbb28-f298-43d3-a416-4aa9ed1f413f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.769667] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 701.769667] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d6d75a-1c29-d561-8cad-df5fa2ce446d" [ 701.769667] env[65788]: _type = "Task" [ 701.769667] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.771784] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290a0f3-3f9f-bbe6-8628-b4e2dd1947df/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 701.771951] env[65788]: ERROR oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290a0f3-3f9f-bbe6-8628-b4e2dd1947df/disk-0.vmdk due to incomplete transfer. [ 701.775497] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1deb21f0-4ea9-490b-b755-1c8786228369 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.784104] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d6d75a-1c29-d561-8cad-df5fa2ce446d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.785813] env[65788]: DEBUG oslo_vmware.rw_handles [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5290a0f3-3f9f-bbe6-8628-b4e2dd1947df/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 701.786068] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Uploaded image d238e3ba-e173-4cb3-b3f0-97cf0daf02f4 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 701.787775] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 701.787963] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-64c4eed7-ed95-408d-a549-e1791af28421 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.795444] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 701.795444] env[65788]: value = "task-4661990" [ 701.795444] env[65788]: _type = "Task" [ 701.795444] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.807341] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661990, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.960484] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Releasing lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.960987] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Instance network_info: |[{"id": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "address": "fa:16:3e:05:d7:71", "network": {"id": "ef52470a-8722-4701-88f2-cb7ecd27ef6b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-878922047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c963c936cf20402d9a65f5e61a11020a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcafb23d-cf", "ovs_interfaceid": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 701.961541] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:d7:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '101a44fc-ffde-4e3e-ad82-363454ae458b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcafb23d-cf58-408a-86ee-d14ae37c1bda', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 701.969017] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Creating folder: Project (c963c936cf20402d9a65f5e61a11020a). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 701.969365] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2cf9a1cf-88b9-4d64-bb0a-db67b9437094 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.982193] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 701.989089] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Created folder: Project (c963c936cf20402d9a65f5e61a11020a) in parent group-v910111. [ 701.989171] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Creating folder: Instances. Parent ref: group-v910205. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 701.989682] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-943a071a-6ae7-424b-a1f1-8c77fa7cdb59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.006044] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Created folder: Instances in parent group-v910205. [ 702.006555] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 702.009016] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 702.009634] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34739e71-ad51-4a90-a8a4-3c146b67888c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.027966] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 702.028358] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 702.028462] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 702.028594] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 702.028776] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 702.028887] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 702.029113] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.029290] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 702.029456] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 702.029632] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 702.029823] env[65788]: DEBUG nova.virt.hardware [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 702.031058] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfa9b74-2227-4634-bb8d-f410b02a4815 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.045266] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf114f3-3be0-4161-9c44-cf50b2df17e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.049779] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 702.049779] env[65788]: value = "task-4661993" [ 702.049779] env[65788]: _type = "Task" [ 702.049779] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.069448] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661993, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.132878] env[65788]: DEBUG nova.scheduler.client.report [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 702.230998] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "6be541bd-06fd-4959-9987-d250011b3f1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.230998] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.269029] env[65788]: DEBUG nova.compute.manager [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 702.269029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 702.269029] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032f40dc-24bc-4fd2-966e-624f0835fb0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.288107] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 702.293216] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1155467-8d26-41d1-84a7-7f39f3c23753 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.296113] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d6d75a-1c29-d561-8cad-df5fa2ce446d, 'name': SearchDatastore_Task, 'duration_secs': 0.03601} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.297365] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.298207] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 702.303796] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.304124] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 702.304617] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89fa6915-8181-40a4-b3ce-5e3ef8dd651d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.308028] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02bd1dc8-a65e-40e8-8346-fb40b8bd52ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.313769] env[65788]: DEBUG oslo_vmware.api [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 702.313769] env[65788]: value = "task-4661994" [ 702.313769] env[65788]: _type = "Task" [ 702.313769] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.324161] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661990, 'name': Destroy_Task, 'duration_secs': 0.431963} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.324567] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 702.324567] env[65788]: value = "task-4661995" [ 702.324567] env[65788]: _type = "Task" [ 702.324567] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.326892] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroyed the VM [ 702.330029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 702.330029] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 702.330029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 702.332909] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1ee79b94-7d7c-433b-b28a-747ac7c40f4a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.339732] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78667787-43e2-42a0-a432-65722b66d10c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.343042] env[65788]: DEBUG oslo_vmware.api [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.347781] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 702.347781] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5283e4f1-97c6-abf1-28de-32baceee30e6" [ 702.347781] env[65788]: _type = "Task" [ 702.347781] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.352239] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.359077] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 702.359077] env[65788]: value = "task-4661996" [ 702.359077] env[65788]: _type = "Task" [ 702.359077] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.368343] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5283e4f1-97c6-abf1-28de-32baceee30e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.375803] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661996, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.562235] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661993, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.638531] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.688s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.639136] env[65788]: DEBUG nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 702.642764] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 25.950s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.684539] env[65788]: DEBUG nova.network.neutron [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Successfully updated port: 83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 702.830247] env[65788]: DEBUG oslo_vmware.api [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661994, 'name': PowerOffVM_Task, 'duration_secs': 0.26758} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.834865] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 702.835119] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 702.836253] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c14e23d-bf00-463a-a4ec-b3b1982c46b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.849311] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661995, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.874032] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5283e4f1-97c6-abf1-28de-32baceee30e6, 'name': SearchDatastore_Task, 'duration_secs': 0.022146} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.874668] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15bbfc6b-f475-4bac-aaa7-68bba831afb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.881271] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661996, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.887081] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 702.887081] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52633e69-40d5-961a-55d3-06f91eeb1e1a" [ 702.887081] env[65788]: _type = "Task" [ 702.887081] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.903861] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52633e69-40d5-961a-55d3-06f91eeb1e1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.008790] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.009150] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.026568] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 703.027176] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 703.027176] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Deleting the datastore file [datastore2] 06f8fc6f-428c-4e40-8815-f0c78d27fa38 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.027176] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6933d7fe-b586-41bc-a9f3-d83934d54346 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.037624] env[65788]: DEBUG oslo_vmware.api [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for the task: (returnval){ [ 703.037624] env[65788]: value = "task-4661999" [ 703.037624] env[65788]: _type = "Task" [ 703.037624] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.051386] env[65788]: DEBUG oslo_vmware.api [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.061978] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4661993, 'name': CreateVM_Task, 'duration_secs': 0.67899} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.062206] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 703.062850] env[65788]: WARNING neutronclient.v2_0.client [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.063806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.063806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.063806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 703.064037] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd0915d7-e3dc-4261-ac22-c3dc882f4399 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.070782] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 703.070782] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ddcf47-f77b-97e2-bc32-8bd545102e95" [ 703.070782] env[65788]: _type = "Task" [ 703.070782] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.080580] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ddcf47-f77b-97e2-bc32-8bd545102e95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.150307] env[65788]: DEBUG nova.compute.utils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 703.165596] env[65788]: DEBUG nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 703.168552] env[65788]: DEBUG nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 703.189757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "refresh_cache-86ac6437-511a-4d72-aff0-0325e2d633f6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.189757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "refresh_cache-86ac6437-511a-4d72-aff0-0325e2d633f6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.189757] env[65788]: DEBUG nova.network.neutron [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 703.343188] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4661995, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.726602} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.344029] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 703.344029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 703.344029] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1769bb27-36fb-4943-b7ca-960eaaafa210 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.352510] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 703.352510] env[65788]: value = "task-4662000" [ 703.352510] env[65788]: _type = "Task" [ 703.352510] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.367592] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662000, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.375401] env[65788]: DEBUG oslo_vmware.api [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4661996, 'name': RemoveSnapshot_Task, 'duration_secs': 0.717665} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.375755] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 703.376011] env[65788]: INFO nova.compute.manager [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 14.95 seconds to snapshot the instance on the hypervisor. [ 703.400344] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52633e69-40d5-961a-55d3-06f91eeb1e1a, 'name': SearchDatastore_Task, 'duration_secs': 0.060593} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.401946] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.402309] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 670f59c3-10b0-431f-a8b0-ef6f82a938be/670f59c3-10b0-431f-a8b0-ef6f82a938be.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 703.404073] env[65788]: DEBUG nova.compute.manager [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Received event network-vif-plugged-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 703.404347] env[65788]: DEBUG oslo_concurrency.lockutils [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Acquiring lock "670f59c3-10b0-431f-a8b0-ef6f82a938be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.404603] env[65788]: DEBUG oslo_concurrency.lockutils [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.404871] env[65788]: DEBUG oslo_concurrency.lockutils [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.405141] env[65788]: DEBUG nova.compute.manager [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] No waiting events found dispatching network-vif-plugged-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 703.405403] env[65788]: WARNING nova.compute.manager [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Received unexpected event network-vif-plugged-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 for instance with vm_state building and task_state spawning. [ 703.405675] env[65788]: DEBUG nova.compute.manager [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Received event network-changed-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 703.405994] env[65788]: DEBUG nova.compute.manager [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Refreshing instance network info cache due to event network-changed-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 703.407420] env[65788]: DEBUG oslo_concurrency.lockutils [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Acquiring lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.407420] env[65788]: DEBUG oslo_concurrency.lockutils [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Acquired lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.407420] env[65788]: DEBUG nova.network.neutron [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Refreshing network info cache for port f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 703.407787] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4aad3d55-728c-413c-8d94-b7de3b40b964 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.419486] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 703.419486] env[65788]: value = "task-4662001" [ 703.419486] env[65788]: _type = "Task" [ 703.419486] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.430375] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.508870] env[65788]: DEBUG nova.compute.manager [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Received event network-vif-plugged-fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 703.509384] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Acquiring lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.513025] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 703.513025] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.513025] env[65788]: DEBUG nova.compute.manager [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] No waiting events found dispatching network-vif-plugged-fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 703.513025] env[65788]: WARNING nova.compute.manager [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Received unexpected event network-vif-plugged-fcafb23d-cf58-408a-86ee-d14ae37c1bda for instance with vm_state building and task_state spawning. [ 703.513025] env[65788]: DEBUG nova.compute.manager [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Received event network-changed-fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 703.513192] env[65788]: DEBUG nova.compute.manager [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Refreshing instance network info cache due to event network-changed-fcafb23d-cf58-408a-86ee-d14ae37c1bda. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 703.513192] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Acquiring lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.513192] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Acquired lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.513192] env[65788]: DEBUG nova.network.neutron [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Refreshing network info cache for port fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 703.555078] env[65788]: DEBUG oslo_vmware.api [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Task: {'id': task-4661999, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.498644} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.555375] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 703.555559] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 703.555719] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.555886] env[65788]: INFO nova.compute.manager [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Took 1.29 seconds to destroy the instance on the hypervisor. [ 703.556154] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 703.556369] env[65788]: DEBUG nova.compute.manager [-] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 703.556462] env[65788]: DEBUG nova.network.neutron [-] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 703.556726] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.557294] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.557557] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.582717] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ddcf47-f77b-97e2-bc32-8bd545102e95, 'name': SearchDatastore_Task, 'duration_secs': 0.027817} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.583224] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.583332] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 703.583866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.583866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.584099] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 703.585184] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8e5ba7b-9e00-4253-a24a-6578184b5287 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.596036] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 703.596036] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 703.596759] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66f0de9f-372f-45f4-b2a7-25d7c9cbcbf6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.603663] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 703.603663] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52808118-16e5-328a-d874-fc26ba37eb7d" [ 703.603663] env[65788]: _type = "Task" [ 703.603663] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.615029] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52808118-16e5-328a-d874-fc26ba37eb7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.652070] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.694326] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.694326] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.719570] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.719754] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance eceafff8-7d28-4b9b-ade6-5736d5977efa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.719875] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 5b701040-025c-4246-ad54-f2cf478e998d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.720022] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4d8f46cd-1c36-4e43-8110-66e9c991f28d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 703.720139] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance b5967423-89e1-44eb-b4de-22d09adb51a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.720250] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 77dcec96-50e5-4753-b3cb-c6aec377bce2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.720360] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance ed55713e-8bf1-4960-8309-47498b163d02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.720469] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 5cef20ed-d9f0-4237-ae9d-da401b953904 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.720724] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 78b545b3-1f84-4ea8-b180-f9039f2ea4b1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 703.720977] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d3e2b5f6-20f1-40ac-b92e-e110ee119178 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.720977] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 957f14fc-23f1-4334-b672-b3a75398f716 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 703.721718] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 27ce0779-2fa2-44d2-a424-072183c022a6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 703.721718] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 703.721718] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 42a9c6f2-08f0-4548-9101-685fc3930c7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.721718] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.721718] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 703.721918] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 06f8fc6f-428c-4e40-8815-f0c78d27fa38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.721918] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance bcdf4a61-a2e8-4ca6-92fc-faddf5371290 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.721982] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.722621] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 670f59c3-10b0-431f-a8b0-ef6f82a938be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.722621] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6215ea56-f99c-42f3-a2d5-f4d31952c8e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.722621] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 86ac6437-511a-4d72-aff0-0325e2d633f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.722621] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 0548399d-ec7a-4d0c-b436-9ef93c4e2d37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 703.777015] env[65788]: DEBUG nova.network.neutron [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 703.840585] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.841412] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.865791] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662000, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074439} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.866126] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.867256] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed10dd3-1c19-4d73-8944-f22faf84739c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.898037] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.902748] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfe8a7d1-ddcb-41fb-8b7e-7d514bfb86e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.927979] env[65788]: DEBUG nova.network.neutron [-] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.930140] env[65788]: WARNING neutronclient.v2_0.client [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.931178] env[65788]: WARNING openstack [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.931334] env[65788]: WARNING openstack [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.949859] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 703.949859] env[65788]: value = "task-4662002" [ 703.949859] env[65788]: _type = "Task" [ 703.949859] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.954460] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662001, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.968806] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662002, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.981339] env[65788]: DEBUG nova.compute.manager [None req-3ace343a-c9bf-4fff-9a43-168324bcc077 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Found 2 images (rotation: 2) {{(pid=65788) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 703.997601] env[65788]: WARNING neutronclient.v2_0.client [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 703.998471] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.998853] env[65788]: WARNING openstack [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.013616] env[65788]: WARNING neutronclient.v2_0.client [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.014494] env[65788]: WARNING openstack [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.014844] env[65788]: WARNING openstack [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.049171] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.049532] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.049786] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "42a9c6f2-08f0-4548-9101-685fc3930c7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.050412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.050412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.052948] env[65788]: INFO nova.compute.manager [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Terminating instance [ 704.119324] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52808118-16e5-328a-d874-fc26ba37eb7d, 'name': SearchDatastore_Task, 'duration_secs': 0.022493} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.120189] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2739f61d-4ca5-463e-9f13-ceff01f06181 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.126921] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 704.126921] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d7868e-a84c-1082-a2c9-0cdcc848679f" [ 704.126921] env[65788]: _type = "Task" [ 704.126921] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.136886] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d7868e-a84c-1082-a2c9-0cdcc848679f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.174517] env[65788]: DEBUG nova.network.neutron [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Updating instance_info_cache with network_info: [{"id": "83cd356c-b01d-42de-8d7b-a6b5cde9d80d", "address": "fa:16:3e:71:0c:63", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83cd356c-b0", "ovs_interfaceid": "83cd356c-b01d-42de-8d7b-a6b5cde9d80d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.184711] env[65788]: DEBUG nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 704.189274] env[65788]: WARNING openstack [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.189274] env[65788]: WARNING openstack [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.199822] env[65788]: WARNING openstack [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.200234] env[65788]: WARNING openstack [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.218674] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 704.218976] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 704.219160] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 704.219351] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 704.219497] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 704.219807] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 704.219936] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.219995] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 704.220178] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 704.220344] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 704.220517] env[65788]: DEBUG nova.virt.hardware [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 704.221729] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c709aa9e-ed1b-449e-8311-97a02516e74b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.229246] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1ef08e8f-744d-4939-9433-a4301cd31e77 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 704.236367] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2e95e3-1af0-4e05-adc3-b2c78eb8b196 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.253687] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.260242] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Creating folder: Project (359e353356894e3eb252292f654cc606). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 704.265719] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df9f3edd-b901-4c98-8111-9b7755b707ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.277505] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Created folder: Project (359e353356894e3eb252292f654cc606) in parent group-v910111. [ 704.277734] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Creating folder: Instances. Parent ref: group-v910208. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 704.278064] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f479fba-3359-4aa6-9ab2-3a9663603a94 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.283045] env[65788]: WARNING neutronclient.v2_0.client [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.283706] env[65788]: WARNING openstack [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.284054] env[65788]: WARNING openstack [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.293872] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Created folder: Instances in parent group-v910208. [ 704.294165] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 704.294377] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.294597] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-168c7a32-8bc1-4a95-8c9e-8594a1da52e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.313688] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.313688] env[65788]: value = "task-4662005" [ 704.313688] env[65788]: _type = "Task" [ 704.313688] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.315032] env[65788]: WARNING neutronclient.v2_0.client [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 704.315626] env[65788]: WARNING openstack [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.316091] env[65788]: WARNING openstack [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.334419] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662005, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.391875] env[65788]: DEBUG nova.network.neutron [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updated VIF entry in instance network info cache for port f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 704.392375] env[65788]: DEBUG nova.network.neutron [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updating instance_info_cache with network_info: [{"id": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "address": "fa:16:3e:4d:7b:e2", "network": {"id": "833e9a86-34c6-4e5f-b367-630030497015", "bridge": "br-int", "label": "tempest-ServersTestJSON-1044181598-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "297851f2d1af4507a43948f51d2f088b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24d1ba6-c0", "ovs_interfaceid": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.422133] env[65788]: DEBUG nova.network.neutron [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Updated VIF entry in instance network info cache for port fcafb23d-cf58-408a-86ee-d14ae37c1bda. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 704.422481] env[65788]: DEBUG nova.network.neutron [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Updating instance_info_cache with network_info: [{"id": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "address": "fa:16:3e:05:d7:71", "network": {"id": "ef52470a-8722-4701-88f2-cb7ecd27ef6b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-878922047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c963c936cf20402d9a65f5e61a11020a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcafb23d-cf", "ovs_interfaceid": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.440977] env[65788]: INFO nova.compute.manager [-] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Took 0.88 seconds to deallocate network for instance. [ 704.454606] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662001, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.464075] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662002, 'name': ReconfigVM_Task, 'duration_secs': 0.459447} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.464393] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.465055] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9cdd23b-50db-417c-ad75-e456c34d10ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.473610] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 704.473610] env[65788]: value = "task-4662006" [ 704.473610] env[65788]: _type = "Task" [ 704.473610] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.485202] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662006, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.557993] env[65788]: DEBUG nova.compute.manager [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 704.558285] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 704.559433] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2dca519-45ea-42c6-96e2-46b8b8171d5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.571631] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 704.571974] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95574e6b-be2a-45a1-b9fb-e0204e67ef8b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.640356] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d7868e-a84c-1082-a2c9-0cdcc848679f, 'name': SearchDatastore_Task, 'duration_secs': 0.05623} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.640860] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.641174] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 6215ea56-f99c-42f3-a2d5-f4d31952c8e6/6215ea56-f99c-42f3-a2d5-f4d31952c8e6.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 704.641508] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2147baa-4eb2-49fd-9166-487c9078eaec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.645988] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 704.646232] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 704.646463] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleting the datastore file [datastore2] 42a9c6f2-08f0-4548-9101-685fc3930c7d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 704.646680] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d895283e-c989-4e34-8fb6-f9d95f23a498 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.651941] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 704.651941] env[65788]: value = "task-4662008" [ 704.651941] env[65788]: _type = "Task" [ 704.651941] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.657961] env[65788]: DEBUG oslo_vmware.api [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 704.657961] env[65788]: value = "task-4662009" [ 704.657961] env[65788]: _type = "Task" [ 704.657961] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.667191] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.674229] env[65788]: DEBUG oslo_vmware.api [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662009, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.682939] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "refresh_cache-86ac6437-511a-4d72-aff0-0325e2d633f6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.682939] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Instance network_info: |[{"id": "83cd356c-b01d-42de-8d7b-a6b5cde9d80d", "address": "fa:16:3e:71:0c:63", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83cd356c-b0", "ovs_interfaceid": "83cd356c-b01d-42de-8d7b-a6b5cde9d80d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 704.683162] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:0c:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83cd356c-b01d-42de-8d7b-a6b5cde9d80d', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 704.691219] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 704.691503] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 704.691757] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29d1a86f-15d8-4c73-9808-9aa8aa8de719 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.721278] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 704.721278] env[65788]: value = "task-4662011" [ 704.721278] env[65788]: _type = "Task" [ 704.721278] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.734609] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6aa4b17b-816c-4d84-8f74-a81185f3af65 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 704.736066] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662011, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.839354] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662005, 'name': CreateVM_Task, 'duration_secs': 0.371452} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.839714] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 704.840243] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.840409] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.841077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 704.841365] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fdc6173f-21cb-4f70-bf11-bc652443b0ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.848075] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 704.848075] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cde2d3-0335-f038-ce7a-a2745c40a34a" [ 704.848075] env[65788]: _type = "Task" [ 704.848075] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.858332] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cde2d3-0335-f038-ce7a-a2745c40a34a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.899311] env[65788]: DEBUG oslo_concurrency.lockutils [req-3531a053-52cb-4964-9a74-aebe77d0448e req-347cdb63-6786-4bf9-8b3f-5a4ec8574a98 service nova] Releasing lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.925327] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e800d06-ae10-467d-b849-9f762b4e84cf req-76ee5948-ae87-475e-9823-68411478a169 service nova] Releasing lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.955268] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662001, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.243196} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.955268] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 670f59c3-10b0-431f-a8b0-ef6f82a938be/670f59c3-10b0-431f-a8b0-ef6f82a938be.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 704.955386] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 704.956433] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.956616] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c014893d-1a38-4f00-8341-7106b8b48461 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.964629] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 704.964629] env[65788]: value = "task-4662012" [ 704.964629] env[65788]: _type = "Task" [ 704.964629] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.976724] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662012, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.986343] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662006, 'name': Rename_Task, 'duration_secs': 0.167264} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.986660] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 704.987342] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b5d8218-ca08-42ee-aab0-383479d256af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.995317] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 704.995317] env[65788]: value = "task-4662013" [ 704.995317] env[65788]: _type = "Task" [ 704.995317] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.006915] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662013, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.082478] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "99f2cc13-a559-4379-8332-d432e8bac36b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.082478] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "99f2cc13-a559-4379-8332-d432e8bac36b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.168665] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662008, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.182508] env[65788]: DEBUG oslo_vmware.api [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662009, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264908} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.182508] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 705.182508] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 705.182508] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.182508] env[65788]: INFO nova.compute.manager [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Took 0.62 seconds to destroy the instance on the hypervisor. [ 705.182748] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 705.182948] env[65788]: DEBUG nova.compute.manager [-] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 705.183123] env[65788]: DEBUG nova.network.neutron [-] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 705.183434] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.183949] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.184274] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.227881] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.231423] env[65788]: DEBUG nova.compute.manager [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 705.233565] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e8190b-f881-416e-8393-55bcf9dc789c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.243042] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f604c16d-2a86-40d3-9891-5b33309b3047 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 705.245129] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662011, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.363133] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cde2d3-0335-f038-ce7a-a2745c40a34a, 'name': SearchDatastore_Task, 'duration_secs': 0.011676} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.363331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.364123] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.364123] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.364123] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.364513] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.364513] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2129b584-be0d-48e0-bcee-5d1f3ba87566 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.376792] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.376792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 705.376792] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1989babb-67e9-4ee0-a73d-d9409f0badd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.383819] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 705.383819] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e7b524-b513-4833-c2ec-fbf1cb16deeb" [ 705.383819] env[65788]: _type = "Task" [ 705.383819] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.393610] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e7b524-b513-4833-c2ec-fbf1cb16deeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.476999] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662012, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077611} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.478060] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 705.479817] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce8d76c-973e-4c7b-941b-8f351cffd22c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.504809] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] 670f59c3-10b0-431f-a8b0-ef6f82a938be/670f59c3-10b0-431f-a8b0-ef6f82a938be.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 705.505281] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73847a3f-f7c9-46cf-b424-7528cbb4fb85 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.531094] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662013, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.532560] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 705.532560] env[65788]: value = "task-4662014" [ 705.532560] env[65788]: _type = "Task" [ 705.532560] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.542702] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662014, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.664290] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673336} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.664589] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 6215ea56-f99c-42f3-a2d5-f4d31952c8e6/6215ea56-f99c-42f3-a2d5-f4d31952c8e6.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 705.664839] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 705.665110] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1083ae31-c993-4810-b608-a9afa6e4b0d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.674116] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 705.674116] env[65788]: value = "task-4662015" [ 705.674116] env[65788]: _type = "Task" [ 705.674116] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.687997] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662015, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.733424] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662011, 'name': CreateVM_Task, 'duration_secs': 0.769237} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.733679] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 705.734324] env[65788]: WARNING neutronclient.v2_0.client [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.734716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.734866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.735269] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 705.735462] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0564b80b-f2e6-4e2a-bad4-58723598c3db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.741480] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 705.741480] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524915f0-6b12-ea39-7d24-1c3a40839f5c" [ 705.741480] env[65788]: _type = "Task" [ 705.741480] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.746796] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 54ad3472-0cca-4dca-91f2-8c343f706926 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 705.755846] env[65788]: INFO nova.compute.manager [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] instance snapshotting [ 705.756505] env[65788]: DEBUG nova.objects.instance [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'flavor' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 705.758129] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524915f0-6b12-ea39-7d24-1c3a40839f5c, 'name': SearchDatastore_Task, 'duration_secs': 0.010731} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.758206] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.758410] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 705.758621] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.894345] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e7b524-b513-4833-c2ec-fbf1cb16deeb, 'name': SearchDatastore_Task, 'duration_secs': 0.013037} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.895304] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aa7bfe8-fcc4-46da-896c-efc003b40dab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.902219] env[65788]: DEBUG nova.compute.manager [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Received event network-vif-plugged-83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 705.902381] env[65788]: DEBUG oslo_concurrency.lockutils [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Acquiring lock "86ac6437-511a-4d72-aff0-0325e2d633f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.902703] env[65788]: DEBUG oslo_concurrency.lockutils [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.902896] env[65788]: DEBUG oslo_concurrency.lockutils [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.903413] env[65788]: DEBUG nova.compute.manager [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] No waiting events found dispatching network-vif-plugged-83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 705.903413] env[65788]: WARNING nova.compute.manager [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Received unexpected event network-vif-plugged-83cd356c-b01d-42de-8d7b-a6b5cde9d80d for instance with vm_state building and task_state spawning. [ 705.903576] env[65788]: DEBUG nova.compute.manager [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Received event network-changed-83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 705.903756] env[65788]: DEBUG nova.compute.manager [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Refreshing instance network info cache due to event network-changed-83cd356c-b01d-42de-8d7b-a6b5cde9d80d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 705.903939] env[65788]: DEBUG oslo_concurrency.lockutils [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Acquiring lock "refresh_cache-86ac6437-511a-4d72-aff0-0325e2d633f6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.904088] env[65788]: DEBUG oslo_concurrency.lockutils [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Acquired lock "refresh_cache-86ac6437-511a-4d72-aff0-0325e2d633f6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.904259] env[65788]: DEBUG nova.network.neutron [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Refreshing network info cache for port 83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 705.907271] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 705.907271] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528569ac-cfc5-66ed-845c-2ecb124f63c6" [ 705.907271] env[65788]: _type = "Task" [ 705.907271] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.918079] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528569ac-cfc5-66ed-845c-2ecb124f63c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010812} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.918820] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.919110] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.919381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 705.919564] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 705.919774] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-237cbf59-9bd8-4c9a-bead-39ec6916bdd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.923059] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c8d10db-4c52-4d90-9300-21e20f454e9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.925582] env[65788]: DEBUG nova.compute.manager [req-c3934c2b-9c94-42cc-87c9-20f9aab9e5cd req-2cf9722d-2fc9-46cd-84d5-d83482cfbcfb service nova] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Received event network-vif-deleted-041beec2-3c7a-44fa-9df0-2681c50094b9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 705.936885] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.937231] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.937558] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 705.937558] env[65788]: value = "task-4662016" [ 705.937558] env[65788]: _type = "Task" [ 705.937558] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.937820] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 705.937981] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 705.939418] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8b9b47a-8b32-487e-a4bb-34da9e31e660 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.949993] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 705.949993] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ea3d45-3f2d-9d9b-afa5-93c24ada46f1" [ 705.949993] env[65788]: _type = "Task" [ 705.949993] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.953466] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.963753] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ea3d45-3f2d-9d9b-afa5-93c24ada46f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010409} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.964547] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72b3b412-ecb0-41d5-be47-ab42ba10cab8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.971262] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 705.971262] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5252a372-9387-c08b-67a7-ca877ee9ceaa" [ 705.971262] env[65788]: _type = "Task" [ 705.971262] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.980720] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5252a372-9387-c08b-67a7-ca877ee9ceaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.016392] env[65788]: DEBUG oslo_vmware.api [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662013, 'name': PowerOnVM_Task, 'duration_secs': 0.576761} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.016694] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 706.016961] env[65788]: INFO nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Took 12.16 seconds to spawn the instance on the hypervisor. [ 706.017114] env[65788]: DEBUG nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 706.018119] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a006ae5d-3d62-4461-b98a-55f09d3ae995 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.045523] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662014, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.052039] env[65788]: DEBUG nova.network.neutron [-] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 706.120646] env[65788]: INFO nova.compute.manager [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Rebuilding instance [ 706.187565] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662015, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170699} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.190824] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 706.192429] env[65788]: DEBUG nova.compute.manager [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 706.192654] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20fb84e9-5999-4db8-b4a6-aaddc0c03287 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.196709] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291cc79b-ef20-4901-815a-aac9d4181167 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.230226] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 6215ea56-f99c-42f3-a2d5-f4d31952c8e6/6215ea56-f99c-42f3-a2d5-f4d31952c8e6.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.233604] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-befeb36c-647c-4750-a1c9-b28127b34022 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.252228] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 3f20bd75-98ab-4798-834a-c0ffc3c6146d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 706.263795] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 706.263795] env[65788]: value = "task-4662017" [ 706.263795] env[65788]: _type = "Task" [ 706.263795] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.270471] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525c0aec-27d0-4b3e-a352-f9653b856feb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.296019] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d7dde5-d056-4ee2-92d9-082e2ef4e444 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.301040] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.409326] env[65788]: WARNING neutronclient.v2_0.client [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 706.410124] env[65788]: WARNING openstack [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.410475] env[65788]: WARNING openstack [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.449729] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662016, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.483463] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5252a372-9387-c08b-67a7-ca877ee9ceaa, 'name': SearchDatastore_Task, 'duration_secs': 0.010656} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.483750] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 706.484030] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 86ac6437-511a-4d72-aff0-0325e2d633f6/86ac6437-511a-4d72-aff0-0325e2d633f6.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 706.484313] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1472b738-8f0b-4ef0-8b8c-2a218d254c64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.493353] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 706.493353] env[65788]: value = "task-4662018" [ 706.493353] env[65788]: _type = "Task" [ 706.493353] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.505803] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662018, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.543995] env[65788]: INFO nova.compute.manager [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Took 38.53 seconds to build instance. [ 706.549754] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662014, 'name': ReconfigVM_Task, 'duration_secs': 0.678377} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.550322] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Reconfigured VM instance instance-0000001d to attach disk [datastore2] 670f59c3-10b0-431f-a8b0-ef6f82a938be/670f59c3-10b0-431f-a8b0-ef6f82a938be.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 706.551230] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d459f9d-31c4-403a-ad02-ccf50d3e7fbd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.554973] env[65788]: INFO nova.compute.manager [-] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Took 1.37 seconds to deallocate network for instance. [ 706.563155] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 706.563155] env[65788]: value = "task-4662019" [ 706.563155] env[65788]: _type = "Task" [ 706.563155] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.578381] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662019, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.616437] env[65788]: WARNING openstack [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.616681] env[65788]: WARNING openstack [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.688882] env[65788]: WARNING neutronclient.v2_0.client [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 706.690108] env[65788]: WARNING openstack [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 706.690108] env[65788]: WARNING openstack [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 706.762565] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 72111766-217d-469f-ad92-ca0ce0e67090 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 706.782499] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.810235] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 706.812228] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fba6f134-069a-40c2-8245-6e353c694bf4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.822349] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 706.822349] env[65788]: value = "task-4662020" [ 706.822349] env[65788]: _type = "Task" [ 706.822349] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.833398] env[65788]: DEBUG nova.network.neutron [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Updated VIF entry in instance network info cache for port 83cd356c-b01d-42de-8d7b-a6b5cde9d80d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 706.834104] env[65788]: DEBUG nova.network.neutron [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Updating instance_info_cache with network_info: [{"id": "83cd356c-b01d-42de-8d7b-a6b5cde9d80d", "address": "fa:16:3e:71:0c:63", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83cd356c-b0", "ovs_interfaceid": "83cd356c-b01d-42de-8d7b-a6b5cde9d80d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 706.839220] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.954754] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524811} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.955013] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.955253] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.955525] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93260a32-313b-4b14-9d86-d9d9383bca7f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.966562] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 706.966562] env[65788]: value = "task-4662022" [ 706.966562] env[65788]: _type = "Task" [ 706.966562] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.978977] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662022, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.007695] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662018, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.047054] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ce8c82fc-8cc2-431b-b35e-87f6623f6809 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.046s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.068748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.080873] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662019, 'name': Rename_Task, 'duration_secs': 0.360778} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.081656] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.081656] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ad742b1-a3bf-4bfd-ab9c-9cd71d81d9a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.091674] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 707.091674] env[65788]: value = "task-4662023" [ 707.091674] env[65788]: _type = "Task" [ 707.091674] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.103982] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662023, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.265803] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 707.266411] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46570fc0-d241-4ddd-80af-a160c3225ee0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.272703] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6be541bd-06fd-4959-9987-d250011b3f1d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 707.279444] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662017, 'name': ReconfigVM_Task, 'duration_secs': 0.614902} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.282594] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 6215ea56-f99c-42f3-a2d5-f4d31952c8e6/6215ea56-f99c-42f3-a2d5-f4d31952c8e6.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.282822] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 707.282822] env[65788]: value = "task-4662024" [ 707.282822] env[65788]: _type = "Task" [ 707.282822] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.283219] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17895e52-a389-4554-8831-8d965182288d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.296676] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662024, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.299549] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 707.299549] env[65788]: value = "task-4662025" [ 707.299549] env[65788]: _type = "Task" [ 707.299549] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.313208] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662025, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.331608] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.340895] env[65788]: DEBUG oslo_concurrency.lockutils [req-c71a735a-9d12-45b1-ba8c-8b89ca8ea9c1 req-7fa6a0e4-04c9-485e-b991-85d0bb951531 service nova] Releasing lock "refresh_cache-86ac6437-511a-4d72-aff0-0325e2d633f6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 707.479839] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662022, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095538} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.480157] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.481093] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949abdf3-1533-4837-8cfd-29b88d9f2e59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.503901] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.508652] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6b9ca16-cde1-4e78-b688-0a985212bda1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.533414] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662018, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689406} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.535303] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 86ac6437-511a-4d72-aff0-0325e2d633f6/86ac6437-511a-4d72-aff0-0325e2d633f6.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 707.535654] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 707.536111] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 707.536111] env[65788]: value = "task-4662026" [ 707.536111] env[65788]: _type = "Task" [ 707.536111] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.536343] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59736882-174f-4bbe-8d90-911463972d0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.552370] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 707.555522] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 707.555522] env[65788]: value = "task-4662027" [ 707.555522] env[65788]: _type = "Task" [ 707.555522] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.560880] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662026, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.574183] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662027, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.610416] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662023, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.782627] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f4a490a2-f2b1-4eac-8c1a-a18758583c70 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 707.783600] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 707.783600] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3840MB phys_disk=100GB used_disk=17GB total_vcpus=48 used_vcpus=17 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '17', 'num_vm_active': '11', 'num_task_None': '11', 'num_os_type_None': '17', 'num_proj_a38ce4af9b414c778a69cd3f64f5f6bf': '1', 'io_workload': '5', 'num_proj_ffbe44d63c1d432e97849f15615329e1': '2', 'num_task_image_uploading': '1', 'num_proj_2b70d3c8627449eaa6372ebe3bd90233': '1', 'num_proj_4fa459531885434f9a977d68864b7042': '1', 'num_proj_4f2bada899b74039930165dd21153376': '1', 'num_proj_6dcc0ac928ee42b9a926c98464c4ec95': '1', 'num_proj_06b3dfec7c0144928b4a59dbd27569b7': '3', 'num_vm_stopped': '1', 'num_proj_3e1713c7002a413fb27180469fded83e': '1', 'num_proj_89909e0b9fd9487982d43248f9022481': '1', 'num_task_deleting': '1', 'num_proj_6284c3d61bc947869abc3e3aefb14846': '1', 'num_proj_9674e2a5c86b48db8c865a50331ab846': '1', 'num_vm_building': '5', 'num_task_spawning': '4', 'num_proj_297851f2d1af4507a43948f51d2f088b': '1', 'num_proj_c963c936cf20402d9a65f5e61a11020a': '1', 'num_proj_359e353356894e3eb252292f654cc606': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 707.807084] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662024, 'name': PowerOffVM_Task, 'duration_secs': 0.52371} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.807374] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 707.807642] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.809143] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36795052-8468-4b83-b89c-2d960f27d904 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.817341] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662025, 'name': Rename_Task, 'duration_secs': 0.488319} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.818223] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.818820] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31321592-6109-4b08-ab57-8a85c4a67ed9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.824254] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 707.834939] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d90833bc-05e0-4354-aaa5-ae86be1b4d03 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.838707] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 707.838707] env[65788]: value = "task-4662028" [ 707.838707] env[65788]: _type = "Task" [ 707.838707] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.850083] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662020, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.859637] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662028, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.937285] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 707.937464] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 707.937565] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.937861] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-732f87df-6c65-4b10-a17f-bb54d19bbb7b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.946109] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 707.946109] env[65788]: value = "task-4662030" [ 707.946109] env[65788]: _type = "Task" [ 707.946109] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.957469] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.052494] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662026, 'name': ReconfigVM_Task, 'duration_secs': 0.361838} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.053064] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.053679] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a91a5c60-e0b3-4629-8a83-55400a0bb956 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.071082] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 708.071082] env[65788]: value = "task-4662031" [ 708.071082] env[65788]: _type = "Task" [ 708.071082] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.084231] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662027, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089911} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.084231] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 708.084231] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41146ef1-3a3d-4366-8308-c925d5775b19 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.092382] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.092382] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662031, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.115609] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 86ac6437-511a-4d72-aff0-0325e2d633f6/86ac6437-511a-4d72-aff0-0325e2d633f6.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 708.119232] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44ff7a43-9db8-4b36-bbf4-a2ae4881ce19 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.138214] env[65788]: DEBUG oslo_vmware.api [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662023, 'name': PowerOnVM_Task, 'duration_secs': 0.817286} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.138887] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.139156] env[65788]: INFO nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Took 11.50 seconds to spawn the instance on the hypervisor. [ 708.139367] env[65788]: DEBUG nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 708.140254] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323535dc-ea85-4530-854d-d45eecae8f6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.143312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.143564] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.143745] env[65788]: DEBUG nova.compute.manager [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 708.145482] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e9dd48-65e1-4727-979a-565bbb7ab517 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.153823] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 708.153823] env[65788]: value = "task-4662032" [ 708.153823] env[65788]: _type = "Task" [ 708.153823] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.162999] env[65788]: DEBUG nova.compute.manager [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 708.163673] env[65788]: DEBUG nova.objects.instance [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'flavor' on Instance uuid 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 708.174509] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662032, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.274847] env[65788]: DEBUG nova.compute.manager [req-47704c59-4fab-4ffd-be87-b24ea2faa2b8 req-80325852-0ae7-42a4-a9e5-8eb6425cd4c2 service nova] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Received event network-vif-deleted-290d5ac2-d3c4-4f0d-b549-523d4ef21f1b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 708.337798] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662020, 'name': CreateSnapshot_Task, 'duration_secs': 1.041238} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.338376] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 708.342136] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7645f2f9-a394-4bfe-a15c-673cca2b7dea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.366760] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662028, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.427672] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413b7673-f867-43cd-9a00-d4c256b34a73 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.440550] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea27f26-3eee-431b-a64f-9876e9b5469b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.479625] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc769b4-0d37-4edb-964e-6a0937684143 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.491392] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64eb9427-4801-44f9-a1c9-6a0fccf8bc68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.495901] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254423} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.496180] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.496368] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 708.496542] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.510498] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.581460] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662031, 'name': Rename_Task, 'duration_secs': 0.146976} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.581858] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 708.582142] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b99b07b8-a322-4e77-84bd-16b4b4a38f2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.591167] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 708.591167] env[65788]: value = "task-4662033" [ 708.591167] env[65788]: _type = "Task" [ 708.591167] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.601678] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.670044] env[65788]: INFO nova.compute.manager [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Took 39.82 seconds to build instance. [ 708.678054] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662032, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.857558] env[65788]: DEBUG oslo_vmware.api [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662028, 'name': PowerOnVM_Task, 'duration_secs': 0.671311} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.857959] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.858262] env[65788]: INFO nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Took 9.56 seconds to spawn the instance on the hypervisor. [ 708.858596] env[65788]: DEBUG nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 708.859194] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76a403b-452a-4904-808e-f9255f87e777 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.876691] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 708.877448] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1a902f9d-90cd-4fe2-a470-5b21fd168f96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.889675] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 708.889675] env[65788]: value = "task-4662035" [ 708.889675] env[65788]: _type = "Task" [ 708.889675] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.900259] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662035, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.013965] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.108040] env[65788]: DEBUG oslo_vmware.api [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662033, 'name': PowerOnVM_Task, 'duration_secs': 0.499153} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.108326] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 709.108525] env[65788]: INFO nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Took 4.92 seconds to spawn the instance on the hypervisor. [ 709.108744] env[65788]: DEBUG nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 709.109565] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31dc6d0-a815-4827-8942-bb904c5fae2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.174043] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8ad78328-eb0f-4f72-b55b-d9b40fcb54d0 tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.333s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.174393] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662032, 'name': ReconfigVM_Task, 'duration_secs': 0.626751} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.175747] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 86ac6437-511a-4d72-aff0-0325e2d633f6/86ac6437-511a-4d72-aff0-0325e2d633f6.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 709.177890] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94b4d8e0-31d9-4661-84a5-b47cfa3639a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.189193] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 709.189193] env[65788]: value = "task-4662036" [ 709.189193] env[65788]: _type = "Task" [ 709.189193] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.194914] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 709.196462] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a988647a-78ac-4782-8a90-e290a8024806 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.204143] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662036, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.205825] env[65788]: DEBUG oslo_vmware.api [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 709.205825] env[65788]: value = "task-4662037" [ 709.205825] env[65788]: _type = "Task" [ 709.205825] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.217807] env[65788]: DEBUG oslo_vmware.api [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.388423] env[65788]: INFO nova.compute.manager [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Took 39.81 seconds to build instance. [ 709.401051] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662035, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.522698] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 709.522698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.880s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.522698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.773s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.522698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.524929] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.777s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.525012] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.530710] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.475s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.530923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.534132] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.307s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.535746] env[65788]: INFO nova.compute.claims [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.548806] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 709.548806] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 709.548806] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 709.549025] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 709.549025] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 709.549498] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 709.549498] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.549660] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 709.549770] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 709.549912] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 709.550224] env[65788]: DEBUG nova.virt.hardware [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 709.550973] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20aa8d16-7d9a-404e-a3d4-14fcb51bf1aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.560717] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c75a65-890f-4972-8161-200ddd83f329 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.580744] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:ca:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42a7074e-6547-4c88-96e5-693f66c36882', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 709.588475] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 709.589676] env[65788]: INFO nova.scheduler.client.report [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleted allocations for instance 78b545b3-1f84-4ea8-b180-f9039f2ea4b1 [ 709.592064] env[65788]: INFO nova.scheduler.client.report [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Deleted allocations for instance 27ce0779-2fa2-44d2-a424-072183c022a6 [ 709.593185] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 709.600013] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b54b49b9-68a1-45db-a7dd-f1dbcb800e8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.622312] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 709.622312] env[65788]: value = "task-4662038" [ 709.622312] env[65788]: _type = "Task" [ 709.622312] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.623435] env[65788]: INFO nova.scheduler.client.report [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Deleted allocations for instance 4d8f46cd-1c36-4e43-8110-66e9c991f28d [ 709.639683] env[65788]: INFO nova.compute.manager [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Took 38.38 seconds to build instance. [ 709.644742] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662038, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.679155] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 709.703305] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662036, 'name': Rename_Task, 'duration_secs': 0.29428} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.703305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 709.703305] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a40ee629-45f0-4fd8-9c0a-0948ba4a470d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.713466] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 709.713466] env[65788]: value = "task-4662039" [ 709.713466] env[65788]: _type = "Task" [ 709.713466] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.722845] env[65788]: DEBUG oslo_vmware.api [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662037, 'name': PowerOffVM_Task, 'duration_secs': 0.460171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.725069] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 709.725069] env[65788]: DEBUG nova.compute.manager [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 709.725069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265e19dc-68e8-4535-984a-462f24c08f16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.732542] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.891167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fdffb2e-e40d-45d9-b94f-9ab9f21aab4b tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.328s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.904594] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662035, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.122267] env[65788]: DEBUG oslo_concurrency.lockutils [None req-609cce84-2a66-4689-9206-819fb06ae0aa tempest-DeleteServersAdminTestJSON-1691143419 tempest-DeleteServersAdminTestJSON-1691143419-project-member] Lock "27ce0779-2fa2-44d2-a424-072183c022a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.738s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.123532] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2b8b1c42-01d7-4a0c-ba98-f3df5eabcaea tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "78b545b3-1f84-4ea8-b180-f9039f2ea4b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.250s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.137866] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662038, 'name': CreateVM_Task, 'duration_secs': 0.387947} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.138830] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 710.141433] env[65788]: WARNING neutronclient.v2_0.client [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 710.142791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.142791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.142791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 710.143164] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7b58700-7726-4c54-8a03-794b625022f9 tempest-ServerMetadataNegativeTestJSON-249816056 tempest-ServerMetadataNegativeTestJSON-249816056-project-member] Lock "4d8f46cd-1c36-4e43-8110-66e9c991f28d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.413s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.144065] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-751c1c2b-0499-4d2f-a2ac-3b9c218a9bc4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.146700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7285fb8d-9fd9-4af1-ad2e-244c6559eab1 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.038s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.153065] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 710.153065] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c487f5-45c4-aac3-103b-fbfab6124d32" [ 710.153065] env[65788]: _type = "Task" [ 710.153065] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.163651] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c487f5-45c4-aac3-103b-fbfab6124d32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.210527] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.227457] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662039, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.245833] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1f59e6dc-0f44-405a-9348-3984732b9efa tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.102s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.399319] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 710.408780] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662035, 'name': CloneVM_Task, 'duration_secs': 1.497961} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.409113] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Created linked-clone VM from snapshot [ 710.409942] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55747da3-54b9-4d48-bb62-835994cad21d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.424669] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Uploading image 4655d41d-1f3e-482a-bce8-16fb2c3f0e05 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 710.467761] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 710.467761] env[65788]: value = "vm-910214" [ 710.467761] env[65788]: _type = "VirtualMachine" [ 710.467761] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 710.468461] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0155bbed-df24-4665-8d93-7e486937819d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.479788] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease: (returnval){ [ 710.479788] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5256207b-898d-1e18-aa80-4ee3675bb450" [ 710.479788] env[65788]: _type = "HttpNfcLease" [ 710.479788] env[65788]: } obtained for exporting VM: (result){ [ 710.479788] env[65788]: value = "vm-910214" [ 710.479788] env[65788]: _type = "VirtualMachine" [ 710.479788] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 710.480331] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the lease: (returnval){ [ 710.480331] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5256207b-898d-1e18-aa80-4ee3675bb450" [ 710.480331] env[65788]: _type = "HttpNfcLease" [ 710.480331] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 710.490944] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 710.490944] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5256207b-898d-1e18-aa80-4ee3675bb450" [ 710.490944] env[65788]: _type = "HttpNfcLease" [ 710.490944] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 710.490944] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 710.490944] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5256207b-898d-1e18-aa80-4ee3675bb450" [ 710.490944] env[65788]: _type = "HttpNfcLease" [ 710.490944] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 710.490944] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b4dc2a-6f36-4ef9-88df-e1806730fc26 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.501922] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bea45b-9175-4032-c289-50f7e3fd6402/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 710.502377] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bea45b-9175-4032-c289-50f7e3fd6402/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 710.608369] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-83d473fb-3ac1-4e29-b6e8-1c954f0f66c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.652905] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 710.671885] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c487f5-45c4-aac3-103b-fbfab6124d32, 'name': SearchDatastore_Task, 'duration_secs': 0.016055} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.672339] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 710.672486] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.672721] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.672781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 710.672955] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.673251] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca8aae13-97b0-4478-aadc-72c93190352d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.686874] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.687127] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 710.687952] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-371b9c39-5235-4b18-a16f-9ad87c14bbd3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.696191] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 710.696191] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ef62b-abe3-1e80-a5f8-bed49d01378f" [ 710.696191] env[65788]: _type = "Task" [ 710.696191] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.707220] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ef62b-abe3-1e80-a5f8-bed49d01378f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.729126] env[65788]: DEBUG oslo_vmware.api [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662039, 'name': PowerOnVM_Task, 'duration_secs': 0.589447} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.729940] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 710.731976] env[65788]: INFO nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Took 8.75 seconds to spawn the instance on the hypervisor. [ 710.731976] env[65788]: DEBUG nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 710.733202] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afed9d7b-f64f-46a6-8fd7-5c40b6153236 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.931595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.181647] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.196969] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd92e1f9-5ea1-4b4c-9e85-9a12650e9748 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.215015] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ef62b-abe3-1e80-a5f8-bed49d01378f, 'name': SearchDatastore_Task, 'duration_secs': 0.01291} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.215833] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0523b490-45b0-425f-914f-66bfb9a2c9f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.220467] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e27f30b-dbb2-4eb8-90a1-77987be72b24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.227364] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 711.227364] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c22eb9-5ee2-c582-56ce-b4401ad12111" [ 711.227364] env[65788]: _type = "Task" [ 711.227364] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.270241] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea1e465-4904-4055-b62e-6bcd780cb872 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.273563] env[65788]: INFO nova.compute.manager [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Took 41.42 seconds to build instance. [ 711.287757] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c22eb9-5ee2-c582-56ce-b4401ad12111, 'name': SearchDatastore_Task, 'duration_secs': 0.026192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.290514] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 711.290901] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 711.294355] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8d172ee-8614-4cd9-b6df-7909f015301d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.297879] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba02081-1980-410c-befb-bc4a62612325 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.320679] env[65788]: DEBUG nova.compute.provider_tree [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.325529] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 711.325529] env[65788]: value = "task-4662042" [ 711.325529] env[65788]: _type = "Task" [ 711.325529] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.342072] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662042, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.780892] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6ca6f1f-03ab-4875-b87b-d6555f74bb03 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.943s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.799657] env[65788]: DEBUG nova.compute.manager [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Received event network-changed-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 711.804096] env[65788]: DEBUG nova.compute.manager [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Refreshing instance network info cache due to event network-changed-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 711.804096] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Acquiring lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.804096] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Acquired lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.804096] env[65788]: DEBUG nova.network.neutron [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Refreshing network info cache for port f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 711.826861] env[65788]: DEBUG nova.scheduler.client.report [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.847385] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662042, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.284886] env[65788]: DEBUG nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 712.305884] env[65788]: WARNING neutronclient.v2_0.client [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.309556] env[65788]: WARNING openstack [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.309882] env[65788]: WARNING openstack [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.338509] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.805s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.339250] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 712.342012] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.460s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.342728] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.344506] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.176s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.346170] env[65788]: INFO nova.compute.claims [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.356698] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662042, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653369} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.356698] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 712.357308] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.357716] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56e8fc8b-a5ec-4a67-a903-6c19d31a4ce4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.368230] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 712.368230] env[65788]: value = "task-4662043" [ 712.368230] env[65788]: _type = "Task" [ 712.368230] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.381937] env[65788]: INFO nova.scheduler.client.report [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Deleted allocations for instance 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8 [ 712.397817] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.537349] env[65788]: WARNING openstack [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.537696] env[65788]: WARNING openstack [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.558855] env[65788]: INFO nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Rebuilding instance [ 712.608381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "b5967423-89e1-44eb-b4de-22d09adb51a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.608796] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "b5967423-89e1-44eb-b4de-22d09adb51a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.609036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "b5967423-89e1-44eb-b4de-22d09adb51a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.609228] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "b5967423-89e1-44eb-b4de-22d09adb51a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.609434] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "b5967423-89e1-44eb-b4de-22d09adb51a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.616090] env[65788]: INFO nova.compute.manager [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Terminating instance [ 712.623067] env[65788]: DEBUG nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 712.624116] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e1ec90-0f27-43f6-ade1-019e65c415a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.806172] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.851990] env[65788]: DEBUG nova.compute.utils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 712.856869] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 712.857250] env[65788]: DEBUG nova.network.neutron [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 712.857703] env[65788]: WARNING neutronclient.v2_0.client [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.858055] env[65788]: WARNING neutronclient.v2_0.client [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 712.858756] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.859757] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.880944] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126164} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.881248] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 712.882145] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ef5869-e98b-4007-976f-e32996d82344 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.914374] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 712.915086] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a981dee0-c19b-4fd8-ae35-ff5296724514 tempest-VolumesAdminNegativeTest-1754688451 tempest-VolumesAdminNegativeTest-1754688451-project-member] Lock "2529ef4c-80d8-47b1-a73a-a02eea0ad8d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.592s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.916862] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c09dcf1-2ea5-4179-bfea-2d1976e90b0b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.941377] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 712.941377] env[65788]: value = "task-4662044" [ 712.941377] env[65788]: _type = "Task" [ 712.941377] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.952860] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662044, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.129531] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "refresh_cache-b5967423-89e1-44eb-b4de-22d09adb51a8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.129875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquired lock "refresh_cache-b5967423-89e1-44eb-b4de-22d09adb51a8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.130227] env[65788]: DEBUG nova.network.neutron [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 713.358169] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 713.455764] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662044, 'name': ReconfigVM_Task, 'duration_secs': 0.474959} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.456119] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Reconfigured VM instance instance-0000001b to attach disk [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290/bcdf4a61-a2e8-4ca6-92fc-faddf5371290.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.457045] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-626e25b3-b326-45fa-b662-88a7e9883bef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.471195] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 713.471195] env[65788]: value = "task-4662046" [ 713.471195] env[65788]: _type = "Task" [ 713.471195] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.485381] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662046, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.566163] env[65788]: WARNING neutronclient.v2_0.client [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.566462] env[65788]: WARNING openstack [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.567170] env[65788]: WARNING openstack [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.632732] env[65788]: WARNING neutronclient.v2_0.client [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 713.633496] env[65788]: WARNING openstack [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.633858] env[65788]: WARNING openstack [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.649072] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 713.649884] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82238086-e2a2-43f3-a5f3-d56e6e981556 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.665271] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 713.665271] env[65788]: value = "task-4662047" [ 713.665271] env[65788]: _type = "Task" [ 713.665271] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.684622] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 713.685227] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 713.686388] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878f68c5-588e-4d84-95fa-fe4fb24d20f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.698031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 713.698421] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d298cac9-642a-4737-bb19-1bd8e8a4c40e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.778834] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 713.778834] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 713.778834] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore2] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 713.779567] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98e08e3d-89a9-4f4d-abf0-db27eb8c7dc9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.789301] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 713.789301] env[65788]: value = "task-4662049" [ 713.789301] env[65788]: _type = "Task" [ 713.789301] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.799102] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662049, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.989682] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662046, 'name': Rename_Task, 'duration_secs': 0.239352} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.989982] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 713.990252] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9665f7a3-d03b-45fa-b4c4-9d4997a12d8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.995691] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3623e52c-28b3-48d4-8114-95aa1a5a3db6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.000488] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 714.000488] env[65788]: value = "task-4662050" [ 714.000488] env[65788]: _type = "Task" [ 714.000488] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.008552] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8817a6-06c2-4bf6-8caa-9d0e82f58d6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.018847] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662050, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.057938] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6ebb1d-9983-495d-b350-9ea7a92a3d9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.070924] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0561f508-27f2-4f1e-a715-7e6e70926c0b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.090856] env[65788]: DEBUG nova.compute.provider_tree [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.300716] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662049, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282852} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.301057] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 714.301249] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 714.301424] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.328878] env[65788]: DEBUG nova.policy [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afdd193345c74bcb9e28467267c1fbd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df050bf0673c41f483ccec937a9d3ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 714.372571] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 714.397244] env[65788]: DEBUG nova.network.neutron [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 714.402813] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 714.403210] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 714.403418] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 714.403621] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 714.403774] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 714.403957] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 714.404239] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.404402] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 714.406739] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 714.406739] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 714.406739] env[65788]: DEBUG nova.virt.hardware [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 714.406739] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2a5bf9-b7d0-4e3d-9a0e-597d4244f58f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.416822] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8f68f1-c2c5-440b-8c45-8d91123a39fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.513371] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662050, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.595801] env[65788]: DEBUG nova.scheduler.client.report [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 714.943559] env[65788]: DEBUG nova.network.neutron [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Successfully created port: 619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 715.013792] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662050, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.103966] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.759s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 715.104542] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 715.107297] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.089s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 715.108794] env[65788]: INFO nova.compute.claims [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.196807] env[65788]: DEBUG nova.network.neutron [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updated VIF entry in instance network info cache for port f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 715.197412] env[65788]: DEBUG nova.network.neutron [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updating instance_info_cache with network_info: [{"id": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "address": "fa:16:3e:4d:7b:e2", "network": {"id": "833e9a86-34c6-4e5f-b367-630030497015", "bridge": "br-int", "label": "tempest-ServersTestJSON-1044181598-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "297851f2d1af4507a43948f51d2f088b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32463b6d-4569-4755-8a29-873a028690a7", "external-id": "nsx-vlan-transportzone-349", "segmentation_id": 349, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24d1ba6-c0", "ovs_interfaceid": "f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 715.339404] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 715.339667] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 715.339837] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 715.340324] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 715.340324] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 715.340402] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 715.340603] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.340778] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 715.340952] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 715.341126] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 715.341285] env[65788]: DEBUG nova.virt.hardware [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 715.342208] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5bb842-c976-42ca-b786-11ef29e75ef1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.351343] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ffccbd-f0f6-490b-afcb-cf4e31ae13f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.367715] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:04:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02852d45-877c-40ee-ade8-47cadad62cf6', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 715.380285] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 715.381498] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 715.381498] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5c12ca0-f51a-48a5-8384-887deade0a90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.401734] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 715.401734] env[65788]: value = "task-4662051" [ 715.401734] env[65788]: _type = "Task" [ 715.401734] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.411358] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662051, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.513699] env[65788]: DEBUG oslo_vmware.api [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662050, 'name': PowerOnVM_Task, 'duration_secs': 1.146893} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.513985] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 715.514212] env[65788]: DEBUG nova.compute.manager [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 715.515069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d979e2-1f1a-470e-85df-7887d19d5820 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.614044] env[65788]: DEBUG nova.compute.utils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 715.618079] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 715.618291] env[65788]: DEBUG nova.network.neutron [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 715.618609] env[65788]: WARNING neutronclient.v2_0.client [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 715.618905] env[65788]: WARNING neutronclient.v2_0.client [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 715.619487] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 715.619883] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 715.700603] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ba79abd-c000-4aea-a089-ee3bd97157ae req-29acfd71-fcf4-47c8-9208-6d7e5963573d service nova] Releasing lock "refresh_cache-670f59c3-10b0-431f-a8b0-ef6f82a938be" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.914025] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662051, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.035254] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.119360] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 716.416224] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662051, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.431216] env[65788]: DEBUG nova.network.neutron [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 716.594473] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0f7b08-7a30-4a54-80b3-19b9bece07a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.603873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62acba26-de95-4336-88d5-84e15fdfe401 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.645037] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242672e0-e5fb-4053-bd4a-f039db411fe0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.656540] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8378210-f713-4f25-96a9-65b5d6cc8cd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.674337] env[65788]: DEBUG nova.compute.provider_tree [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.914827] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662051, 'name': CreateVM_Task, 'duration_secs': 1.041958} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.915061] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 716.915659] env[65788]: WARNING neutronclient.v2_0.client [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 716.916092] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.916255] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 716.916569] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 716.917108] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f686f605-7fa3-46bc-9a50-a2b9deabb023 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.924934] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 716.924934] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528f7210-8d92-86ff-10df-93a2a34d58af" [ 716.924934] env[65788]: _type = "Task" [ 716.924934] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.934516] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528f7210-8d92-86ff-10df-93a2a34d58af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.937345] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Releasing lock "refresh_cache-b5967423-89e1-44eb-b4de-22d09adb51a8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.937974] env[65788]: DEBUG nova.compute.manager [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 716.938305] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.939554] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57310e0f-8185-4deb-aa03-fbbc85838b2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.948434] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 716.948742] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbb1e398-2ca0-46eb-a432-de8cad131db6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.958235] env[65788]: DEBUG oslo_vmware.api [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 716.958235] env[65788]: value = "task-4662052" [ 716.958235] env[65788]: _type = "Task" [ 716.958235] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.969837] env[65788]: DEBUG oslo_vmware.api [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4662052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.996394] env[65788]: DEBUG nova.network.neutron [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Successfully updated port: 619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 717.041671] env[65788]: DEBUG nova.policy [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e801551399b24c1e927aa2341cbbdb6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a73f7788e480480d9236d3a511c034f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 717.152777] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 717.178783] env[65788]: DEBUG nova.scheduler.client.report [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 717.195609] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 717.195893] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 717.196803] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 717.197794] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 717.197958] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 717.198177] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 717.198451] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.198633] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 717.198820] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 717.200625] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 717.200625] env[65788]: DEBUG nova.virt.hardware [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 717.200625] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b913d31-099e-474f-a9a1-01f78d6155e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.211691] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a816880e-ef18-49fe-a363-151d30dbab3d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.437334] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528f7210-8d92-86ff-10df-93a2a34d58af, 'name': SearchDatastore_Task, 'duration_secs': 0.013676} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.437774] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 717.438049] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.439248] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.439492] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.439704] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.440011] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f18d4551-39ed-4b5f-a7cf-1f054bdf4e11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.455026] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.455026] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 717.455406] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c89d889-a8c8-45ce-aba6-82ce477f949b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.465804] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 717.465804] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52728003-e1a7-07c8-5813-cefc3e69e4a5" [ 717.465804] env[65788]: _type = "Task" [ 717.465804] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.473601] env[65788]: DEBUG oslo_vmware.api [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4662052, 'name': PowerOffVM_Task, 'duration_secs': 0.323399} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.474303] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 717.474522] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 717.474851] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a9ded60-8770-4e00-a41a-3fa06222971c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.480305] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52728003-e1a7-07c8-5813-cefc3e69e4a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.504334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "refresh_cache-1ef08e8f-744d-4939-9433-a4301cd31e77" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.504334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired lock "refresh_cache-1ef08e8f-744d-4939-9433-a4301cd31e77" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.504334] env[65788]: DEBUG nova.network.neutron [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 717.508235] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 717.508447] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 717.508629] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleting the datastore file [datastore2] b5967423-89e1-44eb-b4de-22d09adb51a8 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 717.509199] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01b038c0-c521-478a-9e04-fd7eb004e73e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.524032] env[65788]: DEBUG oslo_vmware.api [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for the task: (returnval){ [ 717.524032] env[65788]: value = "task-4662054" [ 717.524032] env[65788]: _type = "Task" [ 717.524032] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.541327] env[65788]: DEBUG oslo_vmware.api [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4662054, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.688023] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.688023] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 717.692393] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.939s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.692393] env[65788]: INFO nova.compute.claims [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.874907] env[65788]: DEBUG nova.network.neutron [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Successfully created port: 97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 717.980087] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52728003-e1a7-07c8-5813-cefc3e69e4a5, 'name': SearchDatastore_Task, 'duration_secs': 0.030741} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.981218] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7125b20a-8829-4a39-a780-93f23b5c7400 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.987619] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 717.987619] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5214414e-a3b5-2cc2-7d8d-3e496aa3c43a" [ 717.987619] env[65788]: _type = "Task" [ 717.987619] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.997554] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5214414e-a3b5-2cc2-7d8d-3e496aa3c43a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.010796] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.011492] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.041435] env[65788]: DEBUG oslo_vmware.api [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Task: {'id': task-4662054, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205235} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.042039] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 718.042039] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 718.042191] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.042385] env[65788]: INFO nova.compute.manager [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Took 1.10 seconds to destroy the instance on the hypervisor. [ 718.042662] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 718.042920] env[65788]: DEBUG nova.compute.manager [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 718.043071] env[65788]: DEBUG nova.network.neutron [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 718.043347] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.043962] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.044606] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.198209] env[65788]: DEBUG nova.compute.utils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 718.203601] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 718.205733] env[65788]: DEBUG nova.network.neutron [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 718.205733] env[65788]: WARNING neutronclient.v2_0.client [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.205733] env[65788]: WARNING neutronclient.v2_0.client [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.206033] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.206480] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.363954] env[65788]: DEBUG nova.network.neutron [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 718.377664] env[65788]: DEBUG nova.network.neutron [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 718.377664] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 718.447855] env[65788]: DEBUG nova.policy [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e441ce452ee48ea8f4ab04c378be68d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89909e0b9fd9487982d43248f9022481', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 718.501970] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5214414e-a3b5-2cc2-7d8d-3e496aa3c43a, 'name': SearchDatastore_Task, 'duration_secs': 0.014703} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.501970] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 718.501970] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 718.501970] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4553611-1d0f-4ae4-85f3-b88ca2851f04 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.509487] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 718.509487] env[65788]: value = "task-4662055" [ 718.509487] env[65788]: _type = "Task" [ 718.509487] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.522850] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662055, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.695544] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.696035] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.717682] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 718.881533] env[65788]: DEBUG nova.network.neutron [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 719.022700] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662055, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.149437] env[65788]: DEBUG nova.network.neutron [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Successfully created port: 325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 719.269197] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bce5a57-a797-4437-974e-90df1481ef90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.279353] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6780bf1d-43d3-4321-8868-2b5511d3696a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.313873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459659c7-a791-4a50-a607-72b7e50cfb90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.322995] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704712f2-ca11-43eb-a3ed-3b4c25daec6d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.339193] env[65788]: DEBUG nova.compute.provider_tree [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.383987] env[65788]: INFO nova.compute.manager [-] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Took 1.34 seconds to deallocate network for instance. [ 719.526579] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662055, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.751717} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.527712] env[65788]: DEBUG nova.network.neutron [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Successfully updated port: 97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 719.528974] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 719.529173] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 719.529412] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd13eb02-3e83-41c2-9f0a-831b4b7a8898 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.538948] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 719.538948] env[65788]: value = "task-4662056" [ 719.538948] env[65788]: _type = "Task" [ 719.538948] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.557261] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662056, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.732593] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 719.769252] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 719.769513] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 719.769665] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 719.769842] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 719.769981] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 719.770141] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 719.770362] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.770533] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 719.770636] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 719.770844] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 719.771060] env[65788]: DEBUG nova.virt.hardware [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 719.771906] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d57b401-cca3-4201-b04c-01e6d4c44257 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.781363] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea49f38d-b86a-4489-b90d-02196bdbda9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.843028] env[65788]: DEBUG nova.scheduler.client.report [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 719.890992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 719.976225] env[65788]: WARNING neutronclient.v2_0.client [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 719.976964] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.977323] env[65788]: WARNING openstack [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.988547] env[65788]: DEBUG nova.compute.manager [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Received event network-changed-fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 719.988737] env[65788]: DEBUG nova.compute.manager [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Refreshing instance network info cache due to event network-changed-fcafb23d-cf58-408a-86ee-d14ae37c1bda. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 719.988959] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Acquiring lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.989165] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Acquired lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.989257] env[65788]: DEBUG nova.network.neutron [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Refreshing network info cache for port fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 720.034777] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "refresh_cache-6aa4b17b-816c-4d84-8f74-a81185f3af65" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.038021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquired lock "refresh_cache-6aa4b17b-816c-4d84-8f74-a81185f3af65" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.038021] env[65788]: DEBUG nova.network.neutron [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 720.056282] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662056, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159076} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.057457] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 720.058629] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fa9727-dbaf-4ac4-bc0f-645d3b54031f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.087024] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 720.087024] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a79514cb-6309-43e9-a248-9816faca2e45 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.111053] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 720.111053] env[65788]: value = "task-4662057" [ 720.111053] env[65788]: _type = "Task" [ 720.111053] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.124345] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662057, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.349973] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.350410] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 720.356539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.562s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.356875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.003s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.359216] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.952s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.360258] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.361822] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.376s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.364033] env[65788]: INFO nova.compute.claims [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.391292] env[65788]: INFO nova.scheduler.client.report [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Deleted allocations for instance d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c [ 720.399045] env[65788]: INFO nova.scheduler.client.report [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Deleted allocations for instance 957f14fc-23f1-4334-b672-b3a75398f716 [ 720.492386] env[65788]: WARNING neutronclient.v2_0.client [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.493181] env[65788]: WARNING openstack [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.493673] env[65788]: WARNING openstack [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.541125] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.541568] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.623108] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662057, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.764294] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "f3c53f67-85bf-4c18-9313-75eb90862f78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 720.764518] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.830038] env[65788]: DEBUG nova.network.neutron [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Updating instance_info_cache with network_info: [{"id": "619462da-ce8b-4b5f-9459-5229925adbb6", "address": "fa:16:3e:67:bd:dc", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap619462da-ce", "ovs_interfaceid": "619462da-ce8b-4b5f-9459-5229925adbb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 720.871059] env[65788]: DEBUG nova.compute.utils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 720.873600] env[65788]: DEBUG nova.network.neutron [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 720.875568] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 720.875928] env[65788]: DEBUG nova.network.neutron [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 720.876485] env[65788]: WARNING neutronclient.v2_0.client [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.877230] env[65788]: WARNING neutronclient.v2_0.client [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 720.877947] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.878467] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.887776] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 720.900270] env[65788]: DEBUG oslo_concurrency.lockutils [None req-870b8f98-3dce-4028-8a8d-e9a7157d91b6 tempest-ServerGroupTestJSON-559816597 tempest-ServerGroupTestJSON-559816597-project-member] Lock "d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.886s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.906510] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a3ae059-b8d3-44aa-b2a1-be1b712286a1 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600 tempest-FloatingIPsAssociationNegativeTestJSON-2090197600-project-member] Lock "957f14fc-23f1-4334-b672-b3a75398f716" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.078s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.124488] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662057, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.178272] env[65788]: DEBUG nova.network.neutron [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Successfully updated port: 325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 721.259074] env[65788]: DEBUG nova.policy [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8ab7f7f29a0418cacca3d10d71829d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a5e0d1546dd64e18a7ddf89faa3562dc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 721.264075] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.264203] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.324888] env[65788]: WARNING openstack [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.325780] env[65788]: WARNING openstack [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.335090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Releasing lock "refresh_cache-1ef08e8f-744d-4939-9433-a4301cd31e77" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.335619] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Instance network_info: |[{"id": "619462da-ce8b-4b5f-9459-5229925adbb6", "address": "fa:16:3e:67:bd:dc", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap619462da-ce", "ovs_interfaceid": "619462da-ce8b-4b5f-9459-5229925adbb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 721.336455] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:bd:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '619462da-ce8b-4b5f-9459-5229925adbb6', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.346200] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Creating folder: Project (df050bf0673c41f483ccec937a9d3ab7). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 721.351641] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a419418c-4fea-4803-ad59-945f43b69817 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.367634] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Created folder: Project (df050bf0673c41f483ccec937a9d3ab7) in parent group-v910111. [ 721.367850] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Creating folder: Instances. Parent ref: group-v910217. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 721.368315] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-066877ef-8c52-42dd-8cc4-8c9cf8ccb120 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.384166] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Created folder: Instances in parent group-v910217. [ 721.384524] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 721.384757] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 721.384991] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9494bbdf-66fc-4bb9-a7e1-36360ef21cd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.419191] env[65788]: INFO nova.virt.block_device [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Booting with volume 54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea at /dev/sda [ 721.422688] env[65788]: WARNING neutronclient.v2_0.client [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.424024] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.424024] env[65788]: WARNING openstack [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.452720] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.452720] env[65788]: value = "task-4662060" [ 721.452720] env[65788]: _type = "Task" [ 721.452720] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.466497] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662060, 'name': CreateVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.516518] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74fb51df-a67c-4d95-892e-7c2d200b3b47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.537229] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135af1d2-c6d0-44ec-baed-9b689c19e951 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.582984] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2eab7ee-35e3-46d2-9451-1468e38e1fc3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.594065] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe83eddf-4cb9-48d5-b40a-b07c2756d537 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.639018] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662057, 'name': ReconfigVM_Task, 'duration_secs': 1.311195} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.644195] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c/4b7f1c2f-2b1e-4d24-814b-c8095d875e3c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 721.645743] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7e2545-c99d-4497-a511-05f4b333732a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.648604] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c0b83b1-c1b0-4cf8-b820-0e444617e8b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.654432] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bea45b-9175-4032-c289-50f7e3fd6402/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 721.657581] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf13fa91-bc88-4d69-bcf3-227f80f33406 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.662817] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 721.662817] env[65788]: value = "task-4662061" [ 721.662817] env[65788]: _type = "Task" [ 721.662817] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.666459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d74558-04af-4b48-aea2-8f155b80a4e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.673522] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bea45b-9175-4032-c289-50f7e3fd6402/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 721.673679] env[65788]: ERROR oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bea45b-9175-4032-c289-50f7e3fd6402/disk-0.vmdk due to incomplete transfer. [ 721.674607] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1a94e233-2e8f-46f3-9981-f00bbbaa41af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.683108] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.683331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.683559] env[65788]: DEBUG nova.network.neutron [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 721.684739] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662061, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.693309] env[65788]: DEBUG oslo_vmware.rw_handles [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bea45b-9175-4032-c289-50f7e3fd6402/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 721.693582] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Uploaded image 4655d41d-1f3e-482a-bce8-16fb2c3f0e05 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 721.696313] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 721.696492] env[65788]: DEBUG nova.virt.block_device [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updating existing volume attachment record: 443a8294-f0ff-43b5-9fcd-5f45439435a8 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 721.699564] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6657b8d8-e2e0-458d-b900-8a973d96f9dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.708369] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 721.708369] env[65788]: value = "task-4662062" [ 721.708369] env[65788]: _type = "Task" [ 721.708369] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.725351] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662062, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.789508] env[65788]: DEBUG nova.network.neutron [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Successfully created port: 2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 721.832969] env[65788]: WARNING neutronclient.v2_0.client [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.833094] env[65788]: WARNING openstack [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.833528] env[65788]: WARNING openstack [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.955476] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b9fc3f-395e-42c9-b0cc-61856873c895 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.970841] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662060, 'name': CreateVM_Task, 'duration_secs': 0.491148} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.971598] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 721.972371] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d01e36bb-afdf-46d4-87fa-49e36d9f3b23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.976457] env[65788]: WARNING neutronclient.v2_0.client [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 721.976857] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.977698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.977698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 721.977698] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82633322-c7f4-4499-a169-17fd4b653a3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.983990] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 721.983990] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f58b5-cf04-991d-d455-12962895334d" [ 721.983990] env[65788]: _type = "Task" [ 721.983990] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.018859] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56bf1e8-faf7-4ff5-bb03-deddec0e7d8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.031562] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a426cb9d-ac1f-43d9-9488-fb3619d90086 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.036379] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f58b5-cf04-991d-d455-12962895334d, 'name': SearchDatastore_Task, 'duration_secs': 0.017788} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.036379] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.036627] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 722.036812] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.036991] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.037551] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 722.038194] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32ac4403-c844-433c-a077-7cf1bb9510f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.051642] env[65788]: DEBUG nova.compute.provider_tree [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.055642] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 722.055828] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 722.056605] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f367881a-7530-4b3c-9f51-c45d65c1dd94 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.063582] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 722.063582] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52814608-7a61-6a0f-5b0f-5ce002fbe4a8" [ 722.063582] env[65788]: _type = "Task" [ 722.063582] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.074521] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52814608-7a61-6a0f-5b0f-5ce002fbe4a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.177333] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662061, 'name': Rename_Task, 'duration_secs': 0.279014} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.177796] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 722.177938] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-935d1d98-90ba-469a-87f5-2a4fea584a62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.186231] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 722.186231] env[65788]: value = "task-4662063" [ 722.186231] env[65788]: _type = "Task" [ 722.186231] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.190388] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 722.190845] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 722.208122] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.220547] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662062, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.381704] env[65788]: DEBUG nova.network.neutron [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Updating instance_info_cache with network_info: [{"id": "97cfb515-1e9e-4a23-bef7-86b855e9ab3f", "address": "fa:16:3e:4d:f0:aa", "network": {"id": "ec8f71e4-924c-45f0-a35a-e77e817f5902", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1745124480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a73f7788e480480d9236d3a511c034f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97cfb515-1e", "ovs_interfaceid": "97cfb515-1e9e-4a23-bef7-86b855e9ab3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 722.556160] env[65788]: DEBUG nova.scheduler.client.report [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 722.576241] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52814608-7a61-6a0f-5b0f-5ce002fbe4a8, 'name': SearchDatastore_Task, 'duration_secs': 0.01091} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.577084] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9cc6955-456a-4d3d-941d-63b0a12b6369 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.584862] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 722.584862] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eb763b-bf27-1934-5f57-8450eecf91cb" [ 722.584862] env[65788]: _type = "Task" [ 722.584862] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.595215] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eb763b-bf27-1934-5f57-8450eecf91cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.697692] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662063, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.720446] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662062, 'name': Destroy_Task, 'duration_secs': 0.690119} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.720670] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroyed the VM [ 722.721094] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 722.721363] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ed88fdff-c6f7-485b-9ad9-ddd0dc855e96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.728360] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 722.728360] env[65788]: value = "task-4662064" [ 722.728360] env[65788]: _type = "Task" [ 722.728360] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.737309] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662064, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.884860] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Releasing lock "refresh_cache-6aa4b17b-816c-4d84-8f74-a81185f3af65" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.885279] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Instance network_info: |[{"id": "97cfb515-1e9e-4a23-bef7-86b855e9ab3f", "address": "fa:16:3e:4d:f0:aa", "network": {"id": "ec8f71e4-924c-45f0-a35a-e77e817f5902", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1745124480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a73f7788e480480d9236d3a511c034f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97cfb515-1e", "ovs_interfaceid": "97cfb515-1e9e-4a23-bef7-86b855e9ab3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 722.885759] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:f0:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97cfb515-1e9e-4a23-bef7-86b855e9ab3f', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.895683] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Creating folder: Project (a73f7788e480480d9236d3a511c034f9). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.896114] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c1d0a49-4db5-4c4c-9070-53881aa6090e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.910803] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Created folder: Project (a73f7788e480480d9236d3a511c034f9) in parent group-v910111. [ 722.911071] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Creating folder: Instances. Parent ref: group-v910220. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.911411] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-067d6712-b236-46a3-9b45-5b8572f7a46a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.922723] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Created folder: Instances in parent group-v910220. [ 722.923052] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 722.923272] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.923538] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8eb188cc-a458-42b8-9f63-2ca674ddfa1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.948677] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.948677] env[65788]: value = "task-4662067" [ 722.948677] env[65788]: _type = "Task" [ 722.948677] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.957711] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662067, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.061761] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.700s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.062724] env[65788]: DEBUG nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 723.066564] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.110s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.066876] env[65788]: DEBUG nova.objects.instance [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lazy-loading 'resources' on Instance uuid 06f8fc6f-428c-4e40-8815-f0c78d27fa38 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 723.104979] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eb763b-bf27-1934-5f57-8450eecf91cb, 'name': SearchDatastore_Task, 'duration_secs': 0.011218} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.105831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.107748] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 1ef08e8f-744d-4939-9433-a4301cd31e77/1ef08e8f-744d-4939-9433-a4301cd31e77.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 723.107748] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c7143a4-2975-450e-a1e4-656c308c61d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.116445] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 723.116445] env[65788]: value = "task-4662068" [ 723.116445] env[65788]: _type = "Task" [ 723.116445] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.128537] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662068, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.198166] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662063, 'name': PowerOnVM_Task, 'duration_secs': 0.520939} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.198442] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 723.198774] env[65788]: DEBUG nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 723.199907] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061a920d-84e4-4b3f-b359-c5cd23d5de1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.241834] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662064, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.301021] env[65788]: DEBUG nova.network.neutron [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 723.461659] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662067, 'name': CreateVM_Task, 'duration_secs': 0.447475} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.462471] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.463031] env[65788]: WARNING neutronclient.v2_0.client [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 723.463217] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.463422] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.463817] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 723.464200] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71c25a9f-62a9-4127-8d05-cda5a7ba21f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.471175] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 723.471175] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5245d713-4379-1668-19ad-e691bc1840a1" [ 723.471175] env[65788]: _type = "Task" [ 723.471175] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.481933] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5245d713-4379-1668-19ad-e691bc1840a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.550846] env[65788]: DEBUG nova.network.neutron [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Successfully updated port: 2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 723.571054] env[65788]: DEBUG nova.compute.utils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 723.572537] env[65788]: DEBUG nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 723.593344] env[65788]: DEBUG nova.network.neutron [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Updated VIF entry in instance network info cache for port fcafb23d-cf58-408a-86ee-d14ae37c1bda. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 723.594365] env[65788]: DEBUG nova.network.neutron [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Updating instance_info_cache with network_info: [{"id": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "address": "fa:16:3e:05:d7:71", "network": {"id": "ef52470a-8722-4701-88f2-cb7ecd27ef6b", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-878922047-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c963c936cf20402d9a65f5e61a11020a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "101a44fc-ffde-4e3e-ad82-363454ae458b", "external-id": "nsx-vlan-transportzone-723", "segmentation_id": 723, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcafb23d-cf", "ovs_interfaceid": "fcafb23d-cf58-408a-86ee-d14ae37c1bda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 723.634298] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662068, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.714866] env[65788]: INFO nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] bringing vm to original state: 'stopped' [ 723.741597] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662064, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.857985] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 723.858627] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 723.858961] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 723.859043] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 723.859191] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 723.859352] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 723.859496] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 723.859708] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.859871] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 723.860057] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 723.860232] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 723.860412] env[65788]: DEBUG nova.virt.hardware [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 723.861630] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df22737b-2c5f-4010-86f0-900f9a6aa9e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.871261] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22da336-885a-4c26-93ad-d12aab236072 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.984341] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5245d713-4379-1668-19ad-e691bc1840a1, 'name': SearchDatastore_Task, 'duration_secs': 0.064644} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.986843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.987091] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.987325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.987465] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.987645] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.988316] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f168f77-b4f3-4224-806b-cafc7036b755 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.998931] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.999188] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.999867] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ced529cb-5495-4fa6-a725-bde946bd9320 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.006598] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 724.006598] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b461e-d8f8-0e45-e119-fe55a96a26e8" [ 724.006598] env[65788]: _type = "Task" [ 724.006598] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.020473] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b461e-d8f8-0e45-e119-fe55a96a26e8, 'name': SearchDatastore_Task, 'duration_secs': 0.009854} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.021473] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-219b6958-1a9b-4bd8-b0d0-bff093215532 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.032200] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 724.032200] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52774728-701a-8280-e198-2e11cdef663b" [ 724.032200] env[65788]: _type = "Task" [ 724.032200] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.043309] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52774728-701a-8280-e198-2e11cdef663b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.053924] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquiring lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.054275] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquired lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.054382] env[65788]: DEBUG nova.network.neutron [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 724.076840] env[65788]: DEBUG nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 724.100378] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffc7f878-e59b-4203-81e0-b54988d8e4f3 req-df8ce740-48b1-416b-bf84-ba950e021e2e service nova] Releasing lock "refresh_cache-6215ea56-f99c-42f3-a2d5-f4d31952c8e6" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.135187] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662068, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530992} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.138841] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 1ef08e8f-744d-4939-9433-a4301cd31e77/1ef08e8f-744d-4939-9433-a4301cd31e77.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 724.139084] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.140144] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ec2d276-1f84-45bc-b917-bc488a680ce8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.144532] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.144906] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.160211] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4843df3-6820-47e9-81d1-4e83e12715cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.168394] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 724.168394] env[65788]: value = "task-4662069" [ 724.168394] env[65788]: _type = "Task" [ 724.168394] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.176611] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908e43fa-4841-4a11-812e-44d9b36dcf77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.185116] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662069, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.217376] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd098644-e7bc-43c0-b9fb-1de0904a768d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.229251] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a962616b-6a5c-4ead-9a28-9ca6f342df05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.245269] env[65788]: DEBUG oslo_vmware.api [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662064, 'name': RemoveSnapshot_Task, 'duration_secs': 1.050495} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.254169] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 724.254360] env[65788]: INFO nova.compute.manager [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 17.99 seconds to snapshot the instance on the hypervisor. [ 724.258080] env[65788]: DEBUG nova.compute.provider_tree [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.517531] env[65788]: WARNING neutronclient.v2_0.client [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 724.518232] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.518593] env[65788]: WARNING openstack [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.544019] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52774728-701a-8280-e198-2e11cdef663b, 'name': SearchDatastore_Task, 'duration_secs': 0.010858} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.544321] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.544578] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6aa4b17b-816c-4d84-8f74-a81185f3af65/6aa4b17b-816c-4d84-8f74-a81185f3af65.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.544855] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-caeade4e-fe5a-4858-bd79-24216cee311c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.554260] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 724.554260] env[65788]: value = "task-4662070" [ 724.554260] env[65788]: _type = "Task" [ 724.554260] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.563176] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.563731] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.571091] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.678641] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662069, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130739} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.678947] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 724.679909] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b10de1-9230-4c5e-8365-3df756f56fc2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.707879] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 1ef08e8f-744d-4939-9433-a4301cd31e77/1ef08e8f-744d-4939-9433-a4301cd31e77.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 724.708314] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4893e850-af4a-48dd-8191-68a0175387cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.723414] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.724035] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.724035] env[65788]: DEBUG nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 724.724873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e22d51-202f-48d4-a159-707be411269a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.734058] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 724.734058] env[65788]: value = "task-4662071" [ 724.734058] env[65788]: _type = "Task" [ 724.734058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.735334] env[65788]: DEBUG nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 724.763485] env[65788]: DEBUG nova.scheduler.client.report [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 724.771029] env[65788]: DEBUG nova.network.neutron [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 724.862231] env[65788]: DEBUG nova.compute.manager [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Found 3 images (rotation: 2) {{(pid=65788) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 724.864211] env[65788]: DEBUG nova.compute.manager [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Rotating out 1 backups {{(pid=65788) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5065}} [ 724.864211] env[65788]: DEBUG nova.compute.manager [None req-defdbdbf-c782-4346-92e1-5ac76d212273 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleting image 65b7cb7b-8569-45e3-b201-5d827f11fc08 {{(pid=65788) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5070}} [ 724.901917] env[65788]: DEBUG nova.network.neutron [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [{"id": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "address": "fa:16:3e:ea:fa:21", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap325d6907-21", "ovs_interfaceid": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 724.928792] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.928792] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.070213] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.087398] env[65788]: DEBUG nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 725.136025] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 725.136633] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 725.136711] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 725.136961] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 725.137133] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 725.137283] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 725.137492] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.138071] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 725.138071] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 725.138071] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 725.138256] env[65788]: DEBUG nova.virt.hardware [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 725.139642] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb07b21-ce55-488a-bf37-5d4fe462dfd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.151597] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8474bbc2-9406-4e02-83b2-ff43196354bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.170400] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.176440] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Creating folder: Project (e004abaed7fd4b3faf6bd9d74d5dc023). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 725.178282] env[65788]: WARNING neutronclient.v2_0.client [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.179011] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.179502] env[65788]: WARNING openstack [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.188030] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d09586c-2183-4e87-a8ac-46de9a2c7d1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.202644] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Created folder: Project (e004abaed7fd4b3faf6bd9d74d5dc023) in parent group-v910111. [ 725.202748] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Creating folder: Instances. Parent ref: group-v910223. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 725.203647] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1168cb36-c675-49cd-a596-7aba6f33e71d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.218415] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Created folder: Instances in parent group-v910223. [ 725.218925] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.219255] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 725.220693] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f8f17ae8-647b-4909-ab4f-0579283f0329 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.244306] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.244306] env[65788]: value = "task-4662074" [ 725.244306] env[65788]: _type = "Task" [ 725.244306] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.248884] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 725.249255] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662071, 'name': ReconfigVM_Task, 'duration_secs': 0.407182} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.252704] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e0ada31-dc13-4e8f-8709-96ccf4ac58fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.254396] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 1ef08e8f-744d-4939-9433-a4301cd31e77/1ef08e8f-744d-4939-9433-a4301cd31e77.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.255224] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d685e1d-034f-4a48-9199-85698acb530f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.264187] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662074, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.267102] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 725.267102] env[65788]: value = "task-4662075" [ 725.267102] env[65788]: _type = "Task" [ 725.267102] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.268057] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 725.268057] env[65788]: value = "task-4662076" [ 725.268057] env[65788]: _type = "Task" [ 725.268057] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.277012] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.210s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.283949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.215s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.284355] env[65788]: DEBUG nova.objects.instance [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lazy-loading 'resources' on Instance uuid 42a9c6f2-08f0-4548-9101-685fc3930c7d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 725.289027] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.294449] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662076, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.320065] env[65788]: INFO nova.scheduler.client.report [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Deleted allocations for instance 06f8fc6f-428c-4e40-8815-f0c78d27fa38 [ 725.356050] env[65788]: DEBUG nova.network.neutron [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updating instance_info_cache with network_info: [{"id": "2117f600-f365-489d-9475-bf0afdc8f512", "address": "fa:16:3e:1f:96:8c", "network": {"id": "8b4c23ed-2ee0-4fc1-b3bf-8306e97e5057", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1568181942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5e0d1546dd64e18a7ddf89faa3562dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2117f600-f3", "ovs_interfaceid": "2117f600-f365-489d-9475-bf0afdc8f512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 725.406044] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.406218] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance network_info: |[{"id": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "address": "fa:16:3e:ea:fa:21", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap325d6907-21", "ovs_interfaceid": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 725.406724] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:fa:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '136c3499-9ca0-4f85-903d-1f194aa66ed9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '325d6907-21b3-4ed1-9291-659b2ee9c13d', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.416566] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.417412] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 725.417747] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8c2623a-f522-44a6-884a-5c6fdf0972c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.444519] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.444519] env[65788]: value = "task-4662077" [ 725.444519] env[65788]: _type = "Task" [ 725.444519] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.456977] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662077, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.568434] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662070, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.592323] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "9438ab56-1b4c-4778-a608-de319ab0ee43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.592323] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.605329] env[65788]: DEBUG nova.compute.manager [req-6712c772-f835-43c5-a537-c9726276ce27 req-bc65a5e5-3088-4d63-aea9-0a521bce2c78 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Received event network-vif-plugged-97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 725.607187] env[65788]: DEBUG oslo_concurrency.lockutils [req-6712c772-f835-43c5-a537-c9726276ce27 req-bc65a5e5-3088-4d63-aea9-0a521bce2c78 service nova] Acquiring lock "6aa4b17b-816c-4d84-8f74-a81185f3af65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.607187] env[65788]: DEBUG oslo_concurrency.lockutils [req-6712c772-f835-43c5-a537-c9726276ce27 req-bc65a5e5-3088-4d63-aea9-0a521bce2c78 service nova] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.607187] env[65788]: DEBUG oslo_concurrency.lockutils [req-6712c772-f835-43c5-a537-c9726276ce27 req-bc65a5e5-3088-4d63-aea9-0a521bce2c78 service nova] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.607187] env[65788]: DEBUG nova.compute.manager [req-6712c772-f835-43c5-a537-c9726276ce27 req-bc65a5e5-3088-4d63-aea9-0a521bce2c78 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] No waiting events found dispatching network-vif-plugged-97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 725.607187] env[65788]: WARNING nova.compute.manager [req-6712c772-f835-43c5-a537-c9726276ce27 req-bc65a5e5-3088-4d63-aea9-0a521bce2c78 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Received unexpected event network-vif-plugged-97cfb515-1e9e-4a23-bef7-86b855e9ab3f for instance with vm_state building and task_state spawning. [ 725.762603] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662074, 'name': CreateVM_Task, 'duration_secs': 0.36716} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.762603] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 725.762893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.763080] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.763722] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 725.764204] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2a01f2b-94e3-451e-8d3a-b831a3822077 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.774981] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 725.774981] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a0d3f9-7c4f-eb1f-45fa-a87d6dae0667" [ 725.774981] env[65788]: _type = "Task" [ 725.774981] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.784295] env[65788]: DEBUG oslo_vmware.api [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662075, 'name': PowerOffVM_Task, 'duration_secs': 0.249291} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.785131] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 725.785289] env[65788]: DEBUG nova.compute.manager [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 725.786403] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44a5b58-f973-41cc-adc0-18b9493f838c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.802721] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a0d3f9-7c4f-eb1f-45fa-a87d6dae0667, 'name': SearchDatastore_Task, 'duration_secs': 0.013831} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.802965] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662076, 'name': Rename_Task, 'duration_secs': 0.220462} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.804530] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.804711] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.805182] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.805251] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.806068] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.806068] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 725.806068] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4054559c-f4b1-48e7-94cc-edfe16e807f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.809218] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd049808-e7ec-4892-b11e-a09abf73fbe8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.823410] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.827035] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.830726] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 725.830726] env[65788]: value = "task-4662078" [ 725.830726] env[65788]: _type = "Task" [ 725.830726] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.832428] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4b36b935-7e73-4402-baa1-a9127ec2bd56 tempest-ServersNegativeTestMultiTenantJSON-993199092 tempest-ServersNegativeTestMultiTenantJSON-993199092-project-member] Lock "06f8fc6f-428c-4e40-8815-f0c78d27fa38" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.083s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.833448] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.833675] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 725.838885] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0f59a9f-514c-42de-b1ca-be1c2f8ac9d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.853105] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 725.853105] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526c0524-57a9-fee4-5e1e-6b4cfd4bd128" [ 725.853105] env[65788]: _type = "Task" [ 725.853105] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.853417] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662078, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.864922] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Releasing lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.865516] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Instance network_info: |[{"id": "2117f600-f365-489d-9475-bf0afdc8f512", "address": "fa:16:3e:1f:96:8c", "network": {"id": "8b4c23ed-2ee0-4fc1-b3bf-8306e97e5057", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1568181942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5e0d1546dd64e18a7ddf89faa3562dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2117f600-f3", "ovs_interfaceid": "2117f600-f365-489d-9475-bf0afdc8f512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 725.866323] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526c0524-57a9-fee4-5e1e-6b4cfd4bd128, 'name': SearchDatastore_Task, 'duration_secs': 0.012501} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.869530] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:96:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c979f78-8597-41f8-b1de-995014032689', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2117f600-f365-489d-9475-bf0afdc8f512', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 725.878875] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Creating folder: Project (a5e0d1546dd64e18a7ddf89faa3562dc). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 725.880880] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-37d84468-6104-49f2-95d6-25fe60de89de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.883239] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-822066a0-7699-44ae-87db-7450ac5e6fb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.894030] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 725.894030] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524dfe60-4056-2a47-1840-7cce3660c99f" [ 725.894030] env[65788]: _type = "Task" [ 725.894030] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.907220] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524dfe60-4056-2a47-1840-7cce3660c99f, 'name': SearchDatastore_Task, 'duration_secs': 0.010852} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.909134] env[65788]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 725.909307] env[65788]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65788) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 725.909986] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.910667] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 3f20bd75-98ab-4798-834a-c0ffc3c6146d/3f20bd75-98ab-4798-834a-c0ffc3c6146d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 725.910667] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Folder already exists: Project (a5e0d1546dd64e18a7ddf89faa3562dc). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 725.911348] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Creating folder: Instances. Parent ref: group-v910178. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 725.911348] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1defc7d0-420d-4d9b-a256-c862e2c0d4b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.913667] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7734917f-95f3-46d1-9441-48da75279d4a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.924629] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 725.924629] env[65788]: value = "task-4662081" [ 725.924629] env[65788]: _type = "Task" [ 725.924629] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.929984] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Created folder: Instances in parent group-v910178. [ 725.930487] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 725.931118] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 725.931557] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bb1504d-aff8-4d4e-a6fd-dd9a1a0085b0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.953136] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.971051] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662077, 'name': CreateVM_Task, 'duration_secs': 0.446474} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.971051] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 725.971051] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.971051] env[65788]: value = "task-4662082" [ 725.971051] env[65788]: _type = "Task" [ 725.971051] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.971703] env[65788]: WARNING neutronclient.v2_0.client [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 725.973037] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.973037] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.973037] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 725.973037] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-451b966d-0e96-498b-82ce-b92e1e524803 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.988832] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 725.988832] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a25864-48a4-6055-cfe4-4cf078cd1a06" [ 725.988832] env[65788]: _type = "Task" [ 725.988832] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.993823] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662082, 'name': CreateVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.004638] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a25864-48a4-6055-cfe4-4cf078cd1a06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.066514] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662070, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.218675} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.069735] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6aa4b17b-816c-4d84-8f74-a81185f3af65/6aa4b17b-816c-4d84-8f74-a81185f3af65.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 726.069972] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.070516] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-836cf328-b675-4631-9e6b-226c58ed5aed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.088059] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 726.088059] env[65788]: value = "task-4662083" [ 726.088059] env[65788]: _type = "Task" [ 726.088059] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.103214] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662083, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.332177] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.608s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.342259] env[65788]: DEBUG nova.compute.manager [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Received event network-vif-plugged-619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 726.342259] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Acquiring lock "1ef08e8f-744d-4939-9433-a4301cd31e77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.342259] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.342259] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.342259] env[65788]: DEBUG nova.compute.manager [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] No waiting events found dispatching network-vif-plugged-619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 726.343909] env[65788]: WARNING nova.compute.manager [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Received unexpected event network-vif-plugged-619462da-ce8b-4b5f-9459-5229925adbb6 for instance with vm_state building and task_state spawning. [ 726.343909] env[65788]: DEBUG nova.compute.manager [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Received event network-changed-619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 726.343909] env[65788]: DEBUG nova.compute.manager [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Refreshing instance network info cache due to event network-changed-619462da-ce8b-4b5f-9459-5229925adbb6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 726.343909] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Acquiring lock "refresh_cache-1ef08e8f-744d-4939-9433-a4301cd31e77" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.343909] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Acquired lock "refresh_cache-1ef08e8f-744d-4939-9433-a4301cd31e77" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.344101] env[65788]: DEBUG nova.network.neutron [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Refreshing network info cache for port 619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 726.364568] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662078, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.431599] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7571d686-8549-448a-a853-cead8e7c903c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.443954] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f7863f-1610-48fa-aa0f-fd1ca01f987b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.447430] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662081, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.483383] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b0a470-8f31-457b-81fb-a75547b375f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.493389] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662082, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.500045] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8411b33b-25df-4cef-80b7-16e482664445 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.516284] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a25864-48a4-6055-cfe4-4cf078cd1a06, 'name': SearchDatastore_Task, 'duration_secs': 0.018406} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.531371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.531591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.531837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.532068] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 726.532500] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.533077] env[65788]: DEBUG nova.compute.provider_tree [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.534614] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b5172ce-59d4-4ac0-8326-1ff3a927acc4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.545165] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.545819] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 726.546853] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63d40b19-a233-4c2e-af26-6d19aaf746e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.556284] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 726.556284] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522d7f3c-e54a-03b3-41ba-5887fe6cf0d4" [ 726.556284] env[65788]: _type = "Task" [ 726.556284] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.564150] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522d7f3c-e54a-03b3-41ba-5887fe6cf0d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.597569] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662083, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079605} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.597858] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.598681] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0bab34f-f4e1-45ae-ae4b-75fc285d898f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.624044] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 6aa4b17b-816c-4d84-8f74-a81185f3af65/6aa4b17b-816c-4d84-8f74-a81185f3af65.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.624746] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d719868-f15c-4c3b-9272-1dac2bb35dde {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.645504] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 726.645504] env[65788]: value = "task-4662084" [ 726.645504] env[65788]: _type = "Task" [ 726.645504] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.649529] env[65788]: INFO nova.compute.manager [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Rebuilding instance [ 726.658291] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662084, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.716136] env[65788]: DEBUG nova.compute.manager [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 726.717085] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27521c8d-4bf5-4306-b09f-82cf81ac1c51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.854603] env[65788]: DEBUG oslo_vmware.api [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662078, 'name': PowerOnVM_Task, 'duration_secs': 0.542963} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.855422] env[65788]: WARNING neutronclient.v2_0.client [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.856038] env[65788]: WARNING openstack [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.856767] env[65788]: WARNING openstack [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.863436] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 726.863436] env[65788]: INFO nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Took 12.49 seconds to spawn the instance on the hypervisor. [ 726.864422] env[65788]: DEBUG nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 726.869098] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a682928d-025b-43d8-be2d-367b8fe0e51b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.875698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 726.937209] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.590326} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.937505] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 3f20bd75-98ab-4798-834a-c0ffc3c6146d/3f20bd75-98ab-4798-834a-c0ffc3c6146d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 726.937712] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.938014] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab732565-6337-4a66-94f2-4532f6db3254 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.948054] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 726.948054] env[65788]: value = "task-4662085" [ 726.948054] env[65788]: _type = "Task" [ 726.948054] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.960445] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662085, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.995838] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662082, 'name': CreateVM_Task, 'duration_secs': 0.716053} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.995838] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 726.996509] env[65788]: WARNING neutronclient.v2_0.client [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.997080] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '443a8294-f0ff-43b5-9fcd-5f45439435a8', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910189', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'name': 'volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '54ad3472-0cca-4dca-91f2-8c343f706926', 'attached_at': '', 'detached_at': '', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'serial': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea'}, 'disk_bus': None, 'device_type': None, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65788) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 726.997363] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Root volume attach. Driver type: vmdk {{(pid=65788) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 726.998358] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb237c67-468f-4799-80ef-75d727f0b994 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.012693] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b1987d-a1ae-4329-bf5c-112fdad064d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.020519] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67aa328b-097c-4b2b-91c2-865bbc1ef8fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.029028] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-0a9937da-6b0e-4b8e-ad9d-ce907e40a8fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.040686] env[65788]: DEBUG nova.scheduler.client.report [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 727.044466] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 727.044466] env[65788]: value = "task-4662086" [ 727.044466] env[65788]: _type = "Task" [ 727.044466] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.054610] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662086, 'name': RelocateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.071620] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522d7f3c-e54a-03b3-41ba-5887fe6cf0d4, 'name': SearchDatastore_Task, 'duration_secs': 0.023407} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.072776] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2142565-474f-4ca4-839f-e497e480b987 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.081239] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 727.081239] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524c5a84-6dc7-9192-a0d4-2d66a52d6267" [ 727.081239] env[65788]: _type = "Task" [ 727.081239] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.091613] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524c5a84-6dc7-9192-a0d4-2d66a52d6267, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.126312] env[65788]: WARNING openstack [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.126702] env[65788]: WARNING openstack [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.161484] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662084, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.231892] env[65788]: WARNING neutronclient.v2_0.client [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 727.232597] env[65788]: WARNING openstack [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.232938] env[65788]: WARNING openstack [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.286934] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.287280] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.287567] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.287814] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.288047] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.291136] env[65788]: INFO nova.compute.manager [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Terminating instance [ 727.398046] env[65788]: INFO nova.compute.manager [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Took 43.19 seconds to build instance. [ 727.409096] env[65788]: DEBUG nova.network.neutron [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Updated VIF entry in instance network info cache for port 619462da-ce8b-4b5f-9459-5229925adbb6. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 727.409096] env[65788]: DEBUG nova.network.neutron [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Updating instance_info_cache with network_info: [{"id": "619462da-ce8b-4b5f-9459-5229925adbb6", "address": "fa:16:3e:67:bd:dc", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap619462da-ce", "ovs_interfaceid": "619462da-ce8b-4b5f-9459-5229925adbb6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 727.461226] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662085, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.205866} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.461591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 727.462602] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c17e8d-f985-4da9-8363-a6958768c02f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.499431] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 3f20bd75-98ab-4798-834a-c0ffc3c6146d/3f20bd75-98ab-4798-834a-c0ffc3c6146d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.500293] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-711316a4-7f1f-42b5-abce-8788d5b1fc0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.524819] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 727.524819] env[65788]: value = "task-4662087" [ 727.524819] env[65788]: _type = "Task" [ 727.524819] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.534504] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662087, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.550971] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.267s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.553535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.463s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.555412] env[65788]: INFO nova.compute.claims [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.564963] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662086, 'name': RelocateVM_Task, 'duration_secs': 0.414057} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.566995] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 727.566995] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910189', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'name': 'volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '54ad3472-0cca-4dca-91f2-8c343f706926', 'attached_at': '', 'detached_at': '', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'serial': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 727.566995] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befe1c4e-6b9d-4abf-bf1c-41ea462150f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.589405] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fcd255-4bee-4953-ad73-ce19594816f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.592710] env[65788]: INFO nova.scheduler.client.report [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted allocations for instance 42a9c6f2-08f0-4548-9101-685fc3930c7d [ 727.610914] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524c5a84-6dc7-9192-a0d4-2d66a52d6267, 'name': SearchDatastore_Task, 'duration_secs': 0.036209} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.624967] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea/volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.625824] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.626177] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f604c16d-2a86-40d3-9891-5b33309b3047/f604c16d-2a86-40d3-9891-5b33309b3047.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 727.626988] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11b3b3a6-d554-4dc1-a728-b68c081bb150 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.646459] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e4ceb31-b711-467b-b622-2951065d2666 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.657714] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 727.657714] env[65788]: value = "task-4662088" [ 727.657714] env[65788]: _type = "Task" [ 727.657714] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.663840] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 727.663840] env[65788]: value = "task-4662089" [ 727.663840] env[65788]: _type = "Task" [ 727.663840] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.664604] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662084, 'name': ReconfigVM_Task, 'duration_secs': 0.872155} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.668236] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 6aa4b17b-816c-4d84-8f74-a81185f3af65/6aa4b17b-816c-4d84-8f74-a81185f3af65.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 727.672017] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-accf61ce-7277-4c88-9e29-3a5dbb542bf2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.681614] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662088, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.688023] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 727.688023] env[65788]: value = "task-4662090" [ 727.688023] env[65788]: _type = "Task" [ 727.688023] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.688383] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662089, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.698693] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662090, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.742097] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 727.742415] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6678203-c692-4824-8433-eb38f101e257 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.751623] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 727.751623] env[65788]: value = "task-4662091" [ 727.751623] env[65788]: _type = "Task" [ 727.751623] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.763709] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.796613] env[65788]: DEBUG nova.compute.manager [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 727.796903] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.798069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0091631-b8ee-4b1f-a8b3-f0b46e6d33e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.808633] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 727.809644] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eaf544c4-0137-48a3-be5b-5524af150a70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.820165] env[65788]: DEBUG oslo_vmware.api [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 727.820165] env[65788]: value = "task-4662092" [ 727.820165] env[65788]: _type = "Task" [ 727.820165] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.842311] env[65788]: DEBUG oslo_vmware.api [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662092, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.901669] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17301729-9926-468b-9683-f01f86b11917 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.709s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.912141] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4e00c4c-34bf-4b1d-a291-b0a0458f7093 req-fb095859-fa29-4a27-81eb-5c970e62914e service nova] Releasing lock "refresh_cache-1ef08e8f-744d-4939-9433-a4301cd31e77" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.046249] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662087, 'name': ReconfigVM_Task, 'duration_secs': 0.366861} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.046555] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 3f20bd75-98ab-4798-834a-c0ffc3c6146d/3f20bd75-98ab-4798-834a-c0ffc3c6146d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.047485] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c70005c-23d5-4037-99d8-0afb046b9083 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.057573] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 728.057573] env[65788]: value = "task-4662093" [ 728.057573] env[65788]: _type = "Task" [ 728.057573] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.072406] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662093, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.129859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6e63d72-4b04-4e4e-bd39-47f78ee181c1 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "42a9c6f2-08f0-4548-9101-685fc3930c7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.079s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.187032] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662088, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.190070] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662089, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.202885] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662090, 'name': Rename_Task, 'duration_secs': 0.206764} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.203408] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 728.204148] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e99699e-91fb-4b92-95fc-033479c18858 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.215693] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 728.215693] env[65788]: value = "task-4662094" [ 728.215693] env[65788]: _type = "Task" [ 728.215693] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.226179] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.270166] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662091, 'name': PowerOffVM_Task, 'duration_secs': 0.243604} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.270503] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 728.271674] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.272956] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67474fe2-7efa-407f-b7bc-bcfc7fa74732 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.286020] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 728.286477] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bb35743-55af-41a0-9615-4069362216ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.319346] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 728.320134] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 728.320491] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Deleting the datastore file [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 728.320991] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-445eb016-0843-4890-b192-6a39ae55e791 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.338050] env[65788]: DEBUG oslo_vmware.api [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662092, 'name': PowerOffVM_Task, 'duration_secs': 0.242306} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.339223] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 728.339337] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 728.339628] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 728.339628] env[65788]: value = "task-4662096" [ 728.339628] env[65788]: _type = "Task" [ 728.339628] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.340633] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4255b32b-9e4e-4de4-abeb-e2bcab5d3dc8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.353657] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.408294] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 728.430818] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 728.435031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 728.435031] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore1] bcdf4a61-a2e8-4ca6-92fc-faddf5371290 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 728.435031] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ccf1370-fc40-4248-bb55-f5df3b2476fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.444641] env[65788]: DEBUG oslo_vmware.api [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 728.444641] env[65788]: value = "task-4662098" [ 728.444641] env[65788]: _type = "Task" [ 728.444641] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.459211] env[65788]: DEBUG oslo_vmware.api [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.571278] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662093, 'name': Rename_Task, 'duration_secs': 0.188867} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.572049] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 728.572711] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-351679de-f6e8-486c-9aae-85a4d77630f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.583199] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 728.583199] env[65788]: value = "task-4662099" [ 728.583199] env[65788]: _type = "Task" [ 728.583199] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.592780] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.683472] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662088, 'name': ReconfigVM_Task, 'duration_secs': 0.660642} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.684691] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Reconfigured VM instance instance-00000024 to attach disk [datastore2] volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea/volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.698231] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eae9e2f8-8c16-4561-afa6-5c3bf18d81bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.708783] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662089, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.709785] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f604c16d-2a86-40d3-9891-5b33309b3047/f604c16d-2a86-40d3-9891-5b33309b3047.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 728.709996] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.711669] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85f127e0-e66e-41e9-9ee8-6c1078cd816e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.720701] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 728.720701] env[65788]: value = "task-4662100" [ 728.720701] env[65788]: _type = "Task" [ 728.720701] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.722721] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 728.722721] env[65788]: value = "task-4662101" [ 728.722721] env[65788]: _type = "Task" [ 728.722721] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.739457] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.743840] env[65788]: DEBUG nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Received event network-changed-97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 728.744081] env[65788]: DEBUG nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Refreshing instance network info cache due to event network-changed-97cfb515-1e9e-4a23-bef7-86b855e9ab3f. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 728.745439] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Acquiring lock "refresh_cache-6aa4b17b-816c-4d84-8f74-a81185f3af65" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.745439] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Acquired lock "refresh_cache-6aa4b17b-816c-4d84-8f74-a81185f3af65" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 728.745851] env[65788]: DEBUG nova.network.neutron [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Refreshing network info cache for port 97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 728.762288] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662100, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.769875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.769875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.770785] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.852861] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112181} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.856489] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.856743] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.857145] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.901010] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.901404] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lock "5b701040-025c-4246-ad54-f2cf478e998d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.948711] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.958710] env[65788]: DEBUG oslo_vmware.api [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152406} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.961792] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.962030] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.962030] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.962213] env[65788]: INFO nova.compute.manager [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Took 1.17 seconds to destroy the instance on the hypervisor. [ 728.962517] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 728.962884] env[65788]: DEBUG nova.compute.manager [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 728.962985] env[65788]: DEBUG nova.network.neutron [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 728.963276] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 728.963905] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.964180] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.021420] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.062670] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "f75000e6-f3bf-4951-a541-3aa185339054" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.062931] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.093156] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662099, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.205486] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6062ac3d-a225-44e0-b7e0-c98279408ead {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.216747] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46e7678-69f8-4287-bb50-449c7d690f25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.251422] env[65788]: WARNING neutronclient.v2_0.client [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.252144] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.252483] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.269544] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d144329-404a-4646-ab73-69b186e0bd56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.272441] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.272861] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662100, 'name': ReconfigVM_Task, 'duration_secs': 0.176199} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.274547] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910189', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'name': 'volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '54ad3472-0cca-4dca-91f2-8c343f706926', 'attached_at': '', 'detached_at': '', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'serial': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 729.275922] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f73cbd9-b59d-4c22-80e5-8310c07e84c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.281077] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090597} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.284553] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.284963] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 729.284963] env[65788]: value = "task-4662102" [ 729.284963] env[65788]: _type = "Task" [ 729.284963] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.286325] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b1e958-17e9-4642-8da8-7786282c3dad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.290201] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1dda63d-9751-410d-b434-08f334d53f4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.305579] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662102, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.323775] env[65788]: DEBUG nova.compute.provider_tree [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.335156] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] f604c16d-2a86-40d3-9891-5b33309b3047/f604c16d-2a86-40d3-9891-5b33309b3047.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.338464] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cd2118a-c160-4b7b-bb28-a81b902b7332 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.361279] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 729.361279] env[65788]: value = "task-4662103" [ 729.361279] env[65788]: _type = "Task" [ 729.361279] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.379227] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662103, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.406057] env[65788]: DEBUG nova.compute.utils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 729.441396] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.441886] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.511152] env[65788]: DEBUG nova.compute.manager [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Received event network-vif-plugged-2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 729.511152] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Acquiring lock "54ad3472-0cca-4dca-91f2-8c343f706926-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.511152] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Lock "54ad3472-0cca-4dca-91f2-8c343f706926-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.511152] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Lock "54ad3472-0cca-4dca-91f2-8c343f706926-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.511152] env[65788]: DEBUG nova.compute.manager [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] No waiting events found dispatching network-vif-plugged-2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 729.511471] env[65788]: WARNING nova.compute.manager [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Received unexpected event network-vif-plugged-2117f600-f365-489d-9475-bf0afdc8f512 for instance with vm_state building and task_state spawning. [ 729.511471] env[65788]: DEBUG nova.compute.manager [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Received event network-changed-2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 729.511471] env[65788]: DEBUG nova.compute.manager [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Refreshing instance network info cache due to event network-changed-2117f600-f365-489d-9475-bf0afdc8f512. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 729.511471] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Acquiring lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.511471] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Acquired lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.511614] env[65788]: DEBUG nova.network.neutron [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Refreshing network info cache for port 2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 729.535911] env[65788]: WARNING neutronclient.v2_0.client [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.536614] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.536958] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.593780] env[65788]: DEBUG oslo_vmware.api [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662099, 'name': PowerOnVM_Task, 'duration_secs': 0.58693} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.594359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 729.594571] env[65788]: INFO nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Took 4.51 seconds to spawn the instance on the hypervisor. [ 729.594751] env[65788]: DEBUG nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 729.600469] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ae2e8b-b260-4558-9d95-6627b3cb7667 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.671215] env[65788]: DEBUG nova.network.neutron [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Updated VIF entry in instance network info cache for port 97cfb515-1e9e-4a23-bef7-86b855e9ab3f. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 729.671575] env[65788]: DEBUG nova.network.neutron [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Updating instance_info_cache with network_info: [{"id": "97cfb515-1e9e-4a23-bef7-86b855e9ab3f", "address": "fa:16:3e:4d:f0:aa", "network": {"id": "ec8f71e4-924c-45f0-a35a-e77e817f5902", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1745124480-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a73f7788e480480d9236d3a511c034f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97cfb515-1e", "ovs_interfaceid": "97cfb515-1e9e-4a23-bef7-86b855e9ab3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.731423] env[65788]: DEBUG oslo_vmware.api [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662094, 'name': PowerOnVM_Task, 'duration_secs': 1.282195} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.731832] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 729.732154] env[65788]: INFO nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Took 12.58 seconds to spawn the instance on the hypervisor. [ 729.732321] env[65788]: DEBUG nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 729.733420] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8e8d31-2c53-4407-9f64-bf6d23819f88 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.775606] env[65788]: DEBUG nova.network.neutron [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.800077] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662102, 'name': Rename_Task, 'duration_secs': 0.152168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.800591] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 729.800985] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5f68d6f-0d0c-4db5-a3bc-7ee50c67eed7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.812052] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 729.812052] env[65788]: value = "task-4662104" [ 729.812052] env[65788]: _type = "Task" [ 729.812052] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.821318] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662104, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.837478] env[65788]: DEBUG nova.scheduler.client.report [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 729.877931] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.910568] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 729.910961] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 729.911374] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 729.911570] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 729.912380] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 729.912380] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 729.912380] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.912611] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 729.912920] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 729.913195] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 729.913495] env[65788]: DEBUG nova.virt.hardware [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 729.914310] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lock "5b701040-025c-4246-ad54-f2cf478e998d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.915780] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161406dc-1599-4c4e-a8e5-f7eca1ffc053 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.926901] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9f8be7-8d72-4098-b3ea-3277f224241f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.944142] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 729.949972] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 729.950323] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 729.950560] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0c6d87f-60fc-4783-b897-f4170f08c41b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.969397] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 729.969397] env[65788]: value = "task-4662105" [ 729.969397] env[65788]: _type = "Task" [ 729.969397] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.980558] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662105, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.015866] env[65788]: WARNING neutronclient.v2_0.client [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.015866] env[65788]: WARNING openstack [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.016584] env[65788]: WARNING openstack [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.120845] env[65788]: INFO nova.compute.manager [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Took 30.17 seconds to build instance. [ 730.158916] env[65788]: WARNING openstack [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.159455] env[65788]: WARNING openstack [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.176035] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Releasing lock "refresh_cache-6aa4b17b-816c-4d84-8f74-a81185f3af65" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 730.176205] env[65788]: DEBUG nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Received event network-vif-plugged-325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 730.176424] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Acquiring lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.176636] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.176796] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.176961] env[65788]: DEBUG nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] No waiting events found dispatching network-vif-plugged-325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 730.177144] env[65788]: WARNING nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Received unexpected event network-vif-plugged-325d6907-21b3-4ed1-9291-659b2ee9c13d for instance with vm_state building and task_state spawning. [ 730.177305] env[65788]: DEBUG nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Received event network-changed-325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 730.177454] env[65788]: DEBUG nova.compute.manager [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Refreshing instance network info cache due to event network-changed-325d6907-21b3-4ed1-9291-659b2ee9c13d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 730.177634] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Acquiring lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.177776] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Acquired lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.178014] env[65788]: DEBUG nova.network.neutron [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Refreshing network info cache for port 325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 730.254574] env[65788]: INFO nova.compute.manager [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Took 45.11 seconds to build instance. [ 730.281809] env[65788]: INFO nova.compute.manager [-] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Took 1.32 seconds to deallocate network for instance. [ 730.322572] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662104, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.345178] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.792s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.345864] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 730.350185] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.139s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.351827] env[65788]: INFO nova.compute.claims [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.363198] env[65788]: WARNING neutronclient.v2_0.client [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.363996] env[65788]: WARNING openstack [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.364341] env[65788]: WARNING openstack [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.385783] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662103, 'name': ReconfigVM_Task, 'duration_secs': 0.760146} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.388045] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Reconfigured VM instance instance-00000023 to attach disk [datastore2] f604c16d-2a86-40d3-9891-5b33309b3047/f604c16d-2a86-40d3-9891-5b33309b3047.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.388854] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96ea0920-c69a-418e-9356-57e3838de1c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.398473] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 730.398473] env[65788]: value = "task-4662106" [ 730.398473] env[65788]: _type = "Task" [ 730.398473] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.407894] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662106, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.484051] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662105, 'name': CreateVM_Task, 'duration_secs': 0.362002} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.484271] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 730.484737] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.484922] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 730.485315] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 730.485591] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b55dddd-3109-4652-a9f8-ffc7fe07e29e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.494603] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 730.494603] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52581c21-02b2-3bfd-c569-4a40cf06ce2a" [ 730.494603] env[65788]: _type = "Task" [ 730.494603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.500659] env[65788]: DEBUG nova.network.neutron [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updated VIF entry in instance network info cache for port 2117f600-f365-489d-9475-bf0afdc8f512. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 730.500817] env[65788]: DEBUG nova.network.neutron [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updating instance_info_cache with network_info: [{"id": "2117f600-f365-489d-9475-bf0afdc8f512", "address": "fa:16:3e:1f:96:8c", "network": {"id": "8b4c23ed-2ee0-4fc1-b3bf-8306e97e5057", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1568181942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5e0d1546dd64e18a7ddf89faa3562dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2117f600-f3", "ovs_interfaceid": "2117f600-f365-489d-9475-bf0afdc8f512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 730.505644] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52581c21-02b2-3bfd-c569-4a40cf06ce2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.623476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8cfa6818-9d06-4330-90c8-a000da8c40ad tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.914s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.682262] env[65788]: WARNING neutronclient.v2_0.client [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.684193] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.684429] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.758092] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d5da777-e4ad-4718-bdd2-d832950e9dcd tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.616s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.788426] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.821930] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.822550] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.837444] env[65788]: DEBUG oslo_vmware.api [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662104, 'name': PowerOnVM_Task, 'duration_secs': 0.548291} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.839817] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 730.839817] env[65788]: INFO nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Took 6.98 seconds to spawn the instance on the hypervisor. [ 730.839817] env[65788]: DEBUG nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 730.839994] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48847269-e178-4461-918d-7b8925428258 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.859326] env[65788]: DEBUG nova.compute.utils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 730.862133] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 730.862393] env[65788]: DEBUG nova.network.neutron [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 730.862876] env[65788]: WARNING neutronclient.v2_0.client [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.863930] env[65788]: WARNING neutronclient.v2_0.client [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.864627] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.864976] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.913616] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662106, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.936550] env[65788]: DEBUG nova.policy [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fff631bddbd74ee2bf389f1b18731f6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d2a28f70fdc4b07b9c4f3e9a6d39725', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 730.974264] env[65788]: WARNING neutronclient.v2_0.client [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 730.975011] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.975564] env[65788]: WARNING openstack [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.013210] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffcae16a-8048-42bb-ae4e-659c8cfa70b2 req-6a816484-3dc3-4ea8-9b90-4c97c53d2cb1 service nova] Releasing lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.013596] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52581c21-02b2-3bfd-c569-4a40cf06ce2a, 'name': SearchDatastore_Task, 'duration_secs': 0.01264} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.014052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.015721] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 731.016317] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.016317] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 731.016476] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 731.017370] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.017370] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lock "5b701040-025c-4246-ad54-f2cf478e998d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.017370] env[65788]: INFO nova.compute.manager [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Attaching volume fe7e6d12-0fe6-42b3-847c-eb3860514545 to /dev/sdb [ 731.019525] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27f404fa-1637-4003-98c7-93d0081c0f5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.036301] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 731.036505] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 731.037314] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8810d547-0303-417b-9040-9ba06ab230e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.045972] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 731.045972] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52216bec-8b02-df4b-6f09-e9ee7afdeac7" [ 731.045972] env[65788]: _type = "Task" [ 731.045972] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.058630] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52216bec-8b02-df4b-6f09-e9ee7afdeac7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.069195] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8351fb49-b2dd-4347-9eda-871226f8896a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.076897] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0be9476-4002-4423-8816-c810501616eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.095213] env[65788]: DEBUG nova.virt.block_device [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updating existing volume attachment record: dc614b1e-1d07-4fc3-9010-5cf589d6e581 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 731.129141] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 731.261606] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 731.317437] env[65788]: DEBUG nova.network.neutron [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Successfully created port: 3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 731.359159] env[65788]: INFO nova.compute.manager [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Took 41.63 seconds to build instance. [ 731.362922] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 731.367671] env[65788]: DEBUG nova.network.neutron [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updated VIF entry in instance network info cache for port 325d6907-21b3-4ed1-9291-659b2ee9c13d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 731.367776] env[65788]: DEBUG nova.network.neutron [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [{"id": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "address": "fa:16:3e:ea:fa:21", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap325d6907-21", "ovs_interfaceid": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 731.409654] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662106, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.500852] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266d888a-c1e5-4631-9534-5e812fc585f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.512949] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867683b5-85b8-4d4e-90ce-46334e631720 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.564134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d267214b-986c-420d-8b13-5912b0bf9cae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.572370] env[65788]: DEBUG nova.compute.manager [req-e925d11c-b7bd-49f1-aa73-6bf01c42e73d req-90e5e72c-f250-44d0-8664-e149cd95cf47 service nova] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Received event network-vif-deleted-42a7074e-6547-4c88-96e5-693f66c36882 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 731.580027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8413419-f26d-4a5c-be35-0da3086a1058 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.585422] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52216bec-8b02-df4b-6f09-e9ee7afdeac7, 'name': SearchDatastore_Task, 'duration_secs': 0.028914} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.587313] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25543214-e404-4778-ad1f-8f983d01f3e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.600229] env[65788]: DEBUG nova.compute.provider_tree [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.606669] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 731.606669] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c1eac-5f29-a804-c898-9860bfa622c0" [ 731.606669] env[65788]: _type = "Task" [ 731.606669] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.622676] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c1eac-5f29-a804-c898-9860bfa622c0, 'name': SearchDatastore_Task, 'duration_secs': 0.01226} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.623748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.624062] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 731.624350] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7299328-7cf9-45c1-abbf-0453fe4147b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.637031] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 731.637031] env[65788]: value = "task-4662110" [ 731.637031] env[65788]: _type = "Task" [ 731.637031] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.651401] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662110, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.658034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.702359] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.702723] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.702992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.703261] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 731.703467] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.706419] env[65788]: INFO nova.compute.manager [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Terminating instance [ 731.789412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.861377] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3e8612b8-3e63-4dd2-be2e-3e8fe3ed59d0 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "54ad3472-0cca-4dca-91f2-8c343f706926" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.247s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.875264] env[65788]: DEBUG oslo_concurrency.lockutils [req-f68d9c31-cd54-48aa-8dd2-4b572be3433e req-9d052f77-94ae-4f3a-8570-d5ab3af9e895 service nova] Releasing lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 731.917467] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662106, 'name': Rename_Task, 'duration_secs': 1.32784} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.917824] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 731.918062] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06b9850e-dd7b-443c-89cf-a77b0d421d8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.927094] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 731.927094] env[65788]: value = "task-4662111" [ 731.927094] env[65788]: _type = "Task" [ 731.927094] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.943014] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662111, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.108999] env[65788]: DEBUG nova.scheduler.client.report [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 732.153589] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662110, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.211307] env[65788]: DEBUG nova.compute.manager [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 732.211721] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.212511] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74aec8e1-b2fd-4987-bdc8-d14ffd504569 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.222380] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.222691] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57f1d2ef-936a-44d8-8f92-a896698d28a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.306228] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.306554] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.306768] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore1] 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.307105] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b11b324d-2b6e-466b-b519-71340f411f11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.316294] env[65788]: DEBUG oslo_vmware.api [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 732.316294] env[65788]: value = "task-4662113" [ 732.316294] env[65788]: _type = "Task" [ 732.316294] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.329542] env[65788]: DEBUG oslo_vmware.api [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.364345] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 732.377238] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 732.404501] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 732.404769] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 732.404921] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 732.405152] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 732.405309] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 732.405456] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 732.405712] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.405925] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 732.406137] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 732.406309] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 732.406549] env[65788]: DEBUG nova.virt.hardware [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 732.407576] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4fd9d0b-559d-4fae-b07a-5baf211cfef5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.418177] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a00768-dae0-4225-893c-f8fa472e690f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.445877] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662111, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.523884] env[65788]: DEBUG nova.compute.manager [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 732.524843] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2714c0ff-dc3c-4bb7-a9d4-6f6cf6f9bf53 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.614950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.615715] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 732.619342] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.688s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.622099] env[65788]: INFO nova.compute.claims [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.654611] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662110, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558099} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.654768] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.655239] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.655239] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f29c43f9-243b-46a9-aae6-15aa5faac883 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.667751] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 732.667751] env[65788]: value = "task-4662114" [ 732.667751] env[65788]: _type = "Task" [ 732.667751] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.680815] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662114, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.833313] env[65788]: DEBUG oslo_vmware.api [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456169} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.833313] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.833403] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 732.833677] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.833982] env[65788]: INFO nova.compute.manager [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Took 0.62 seconds to destroy the instance on the hypervisor. [ 732.834426] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 732.837125] env[65788]: DEBUG nova.compute.manager [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 732.837496] env[65788]: DEBUG nova.network.neutron [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 732.837859] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 732.838950] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.839065] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.893159] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.947958] env[65788]: DEBUG oslo_vmware.api [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662111, 'name': PowerOnVM_Task, 'duration_secs': 1.013525} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.948478] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 732.948598] env[65788]: INFO nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Took 13.22 seconds to spawn the instance on the hypervisor. [ 732.948831] env[65788]: DEBUG nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 732.949805] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd26be86-c636-456f-991b-16dd47f00b0b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.043061] env[65788]: INFO nova.compute.manager [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] instance snapshotting [ 733.043960] env[65788]: DEBUG nova.objects.instance [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lazy-loading 'flavor' on Instance uuid 3f20bd75-98ab-4798-834a-c0ffc3c6146d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 733.093650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.093867] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.094561] env[65788]: DEBUG nova.compute.manager [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 733.094945] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496e70c2-5403-4103-813c-928242852ad5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.103583] env[65788]: DEBUG nova.compute.manager [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 733.104198] env[65788]: DEBUG nova.objects.instance [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'flavor' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 733.127601] env[65788]: DEBUG nova.compute.utils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 733.131681] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 733.131930] env[65788]: DEBUG nova.network.neutron [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 733.132322] env[65788]: WARNING neutronclient.v2_0.client [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 733.132639] env[65788]: WARNING neutronclient.v2_0.client [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 733.133283] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.133617] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.178577] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662114, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.17292} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.178947] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 733.179712] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8deebdc4-69a1-4a16-97fc-294ae9589960 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.203533] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.203881] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0755029-a0f6-4cb1-a358-3c1a9b144443 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.226246] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 733.226246] env[65788]: value = "task-4662115" [ 733.226246] env[65788]: _type = "Task" [ 733.226246] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.236971] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662115, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.481739] env[65788]: INFO nova.compute.manager [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Took 47.49 seconds to build instance. [ 733.556139] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51841b4d-c660-4fa4-80ea-8729a4ad3dd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.559939] env[65788]: DEBUG nova.network.neutron [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Successfully updated port: 3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 733.596768] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc276bf-7cfa-4deb-a288-698628a0433f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.627262] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 733.642423] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 733.667409] env[65788]: DEBUG nova.policy [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '708b59a453154170b288b3804c98313a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6002001018ca4f6983504dac34952e0a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 733.741804] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662115, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.985451] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56f0eb6e-6722-4551-9519-31d210f6d0c1 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.068241] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "refresh_cache-72111766-217d-469f-ad92-ca0ce0e67090" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.068241] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquired lock "refresh_cache-72111766-217d-469f-ad92-ca0ce0e67090" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.068241] env[65788]: DEBUG nova.network.neutron [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 734.116127] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 734.116127] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f2712f4c-c3ad-4391-bee4-96711ec76a6e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.121477] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 734.122760] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6f39792-ebbc-4bd2-94d0-c7c4bc7458a1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.132933] env[65788]: DEBUG oslo_vmware.api [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 734.132933] env[65788]: value = "task-4662117" [ 734.132933] env[65788]: _type = "Task" [ 734.132933] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.135825] env[65788]: DEBUG oslo_vmware.api [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 734.135825] env[65788]: value = "task-4662118" [ 734.135825] env[65788]: _type = "Task" [ 734.135825] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.158674] env[65788]: DEBUG oslo_vmware.api [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.159416] env[65788]: DEBUG oslo_vmware.api [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662118, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.247640] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662115, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.277561] env[65788]: DEBUG nova.network.neutron [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Successfully created port: 35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 734.354651] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518cb691-c99b-4fd0-964f-0989d3bb0981 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.365127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb11eae3-48fc-4a1d-9512-a1eb6de75796 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.399390] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71910ff8-268d-41f1-b30a-d8c3e6dfde53 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.408731] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b94d91b-6ce6-46a1-9dfa-8a56438fb782 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.426567] env[65788]: DEBUG nova.compute.provider_tree [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.446102] env[65788]: DEBUG nova.network.neutron [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 734.492468] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 734.499819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "5c7a1693-62f2-454e-9406-0b4a132ebf25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.500146] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.579842] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.580282] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.624872] env[65788]: DEBUG nova.network.neutron [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 734.644327] env[65788]: DEBUG oslo_vmware.api [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662117, 'name': PowerOffVM_Task, 'duration_secs': 0.313789} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.648016] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.648340] env[65788]: DEBUG nova.compute.manager [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 734.649131] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec15fbf-5fe3-4838-9596-442b9b2f73b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.658664] env[65788]: DEBUG oslo_vmware.api [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662118, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.660313] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.660562] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.672739] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 734.717460] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 734.717836] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 734.718090] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 734.718367] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 734.718586] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 734.718810] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 734.719137] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.719387] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 734.719685] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 734.719878] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 734.720158] env[65788]: DEBUG nova.virt.hardware [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 734.722912] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9993bce-94a2-45e2-9cf7-526dba33629f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.740999] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6753fe-1686-45c3-b6db-1b69efe44442 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.753658] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662115, 'name': ReconfigVM_Task, 'duration_secs': 1.073895} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.754996] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37/0548399d-ec7a-4d0c-b436-9ef93c4e2d37.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 734.755696] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e8c9caf-8e8c-4780-952b-330801a7f299 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.783696] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 734.783696] env[65788]: value = "task-4662119" [ 734.783696] env[65788]: _type = "Task" [ 734.783696] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.794713] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662119, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.795958] env[65788]: WARNING neutronclient.v2_0.client [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 734.796799] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.797929] env[65788]: WARNING openstack [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.852796] env[65788]: DEBUG nova.compute.manager [req-95aa705b-8a60-42f3-85d5-639b3a10c5b6 req-05a63237-9f90-4434-af91-54ab46b8c34c service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Received event network-vif-plugged-3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 734.852796] env[65788]: DEBUG oslo_concurrency.lockutils [req-95aa705b-8a60-42f3-85d5-639b3a10c5b6 req-05a63237-9f90-4434-af91-54ab46b8c34c service nova] Acquiring lock "72111766-217d-469f-ad92-ca0ce0e67090-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.853270] env[65788]: DEBUG oslo_concurrency.lockutils [req-95aa705b-8a60-42f3-85d5-639b3a10c5b6 req-05a63237-9f90-4434-af91-54ab46b8c34c service nova] Lock "72111766-217d-469f-ad92-ca0ce0e67090-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.853624] env[65788]: DEBUG oslo_concurrency.lockutils [req-95aa705b-8a60-42f3-85d5-639b3a10c5b6 req-05a63237-9f90-4434-af91-54ab46b8c34c service nova] Lock "72111766-217d-469f-ad92-ca0ce0e67090-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.854094] env[65788]: DEBUG nova.compute.manager [req-95aa705b-8a60-42f3-85d5-639b3a10c5b6 req-05a63237-9f90-4434-af91-54ab46b8c34c service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] No waiting events found dispatching network-vif-plugged-3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 734.854420] env[65788]: WARNING nova.compute.manager [req-95aa705b-8a60-42f3-85d5-639b3a10c5b6 req-05a63237-9f90-4434-af91-54ab46b8c34c service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Received unexpected event network-vif-plugged-3c6f2879-19ae-45b8-8785-45987a597a6a for instance with vm_state building and task_state spawning. [ 734.928570] env[65788]: DEBUG nova.scheduler.client.report [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 734.951425] env[65788]: INFO nova.compute.manager [-] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Took 2.11 seconds to deallocate network for instance. [ 734.969992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.969992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.970146] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.971941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.971941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.973066] env[65788]: INFO nova.compute.manager [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Terminating instance [ 734.991480] env[65788]: DEBUG nova.network.neutron [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Updating instance_info_cache with network_info: [{"id": "3c6f2879-19ae-45b8-8785-45987a597a6a", "address": "fa:16:3e:3b:35:bb", "network": {"id": "934955b9-cb1d-49c3-9a97-76c56f073eab", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-107342658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8d2a28f70fdc4b07b9c4f3e9a6d39725", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c6f2879-19", "ovs_interfaceid": "3c6f2879-19ae-45b8-8785-45987a597a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 735.028229] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.157479] env[65788]: DEBUG oslo_vmware.api [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662118, 'name': CreateSnapshot_Task, 'duration_secs': 0.844884} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.157913] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 735.158597] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a42fce-7590-4f9e-9d94-57013e222be3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.181193] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e99e6b67-bc11-45ba-bfeb-6475a7109b6e tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.087s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.199485] env[65788]: DEBUG nova.compute.manager [req-ce7471ca-7516-4e32-812c-fd6cf9db9e2c req-ecb54340-325e-4981-8e23-cfe7714b1e9c service nova] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Received event network-vif-deleted-02852d45-877c-40ee-ade8-47cadad62cf6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 735.295738] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662119, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.435364] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.816s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.436277] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 735.438777] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.259s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.440201] env[65788]: INFO nova.compute.claims [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.466455] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.478377] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "refresh_cache-3f20bd75-98ab-4798-834a-c0ffc3c6146d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.478552] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquired lock "refresh_cache-3f20bd75-98ab-4798-834a-c0ffc3c6146d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.478774] env[65788]: DEBUG nova.network.neutron [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 735.495616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Releasing lock "refresh_cache-72111766-217d-469f-ad92-ca0ce0e67090" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 735.496117] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Instance network_info: |[{"id": "3c6f2879-19ae-45b8-8785-45987a597a6a", "address": "fa:16:3e:3b:35:bb", "network": {"id": "934955b9-cb1d-49c3-9a97-76c56f073eab", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-107342658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8d2a28f70fdc4b07b9c4f3e9a6d39725", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c6f2879-19", "ovs_interfaceid": "3c6f2879-19ae-45b8-8785-45987a597a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 735.496595] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:35:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3093647a-bab7-4562-ada0-428725e8c0fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c6f2879-19ae-45b8-8785-45987a597a6a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 735.505561] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Creating folder: Project (8d2a28f70fdc4b07b9c4f3e9a6d39725). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 735.506385] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96dc473c-a963-4045-97d3-0398942a663a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.520699] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Created folder: Project (8d2a28f70fdc4b07b9c4f3e9a6d39725) in parent group-v910111. [ 735.520699] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Creating folder: Instances. Parent ref: group-v910234. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 735.520699] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cba0621f-965e-4633-bfcd-02ba7f256891 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.533258] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Created folder: Instances in parent group-v910234. [ 735.533540] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 735.533774] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 735.533774] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9136482f-8e1a-4f08-9c72-01b9f1933aad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.554085] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 735.554085] env[65788]: value = "task-4662122" [ 735.554085] env[65788]: _type = "Task" [ 735.554085] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.564270] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662122, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.671434] env[65788]: DEBUG nova.compute.manager [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Instance disappeared during snapshot {{(pid=65788) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4635}} [ 735.675291] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 735.675291] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910232', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'name': 'volume-fe7e6d12-0fe6-42b3-847c-eb3860514545', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b701040-025c-4246-ad54-f2cf478e998d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'serial': 'fe7e6d12-0fe6-42b3-847c-eb3860514545'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 735.676169] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8362895-39fe-44af-bbea-1ed6652e68bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.699310] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a8f309-c5de-4eca-ae07-3097fec0537b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.731611] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] volume-fe7e6d12-0fe6-42b3-847c-eb3860514545/volume-fe7e6d12-0fe6-42b3-847c-eb3860514545.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.736017] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39fb5396-69a2-4242-8d15-ec3f437560fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.758338] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Waiting for the task: (returnval){ [ 735.758338] env[65788]: value = "task-4662123" [ 735.758338] env[65788]: _type = "Task" [ 735.758338] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.768638] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662123, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.796581] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662119, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.910600] env[65788]: DEBUG nova.network.neutron [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Successfully updated port: 35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 735.917415] env[65788]: DEBUG nova.compute.manager [None req-6e97bdf6-ca28-47db-973b-ebdf55ecd12b tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Found 0 images (rotation: 2) {{(pid=65788) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5057}} [ 735.945791] env[65788]: DEBUG nova.compute.utils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 735.955442] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 735.956349] env[65788]: DEBUG nova.network.neutron [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 735.956349] env[65788]: WARNING neutronclient.v2_0.client [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.956825] env[65788]: WARNING neutronclient.v2_0.client [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.959830] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.959830] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.982487] env[65788]: WARNING neutronclient.v2_0.client [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.983942] env[65788]: WARNING openstack [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.984484] env[65788]: WARNING openstack [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.067579] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662122, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.270626] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662123, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.298149] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662119, 'name': Rename_Task, 'duration_secs': 1.201485} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.298657] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 736.299058] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-958ed2bc-db8b-4d12-8d2e-9fa3332ba3a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.309336] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 736.309336] env[65788]: value = "task-4662124" [ 736.309336] env[65788]: _type = "Task" [ 736.309336] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.332244] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662124, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.414698] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.414889] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquired lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.415372] env[65788]: DEBUG nova.network.neutron [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 736.455717] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 736.505704] env[65788]: DEBUG nova.network.neutron [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 736.564103] env[65788]: DEBUG nova.policy [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bedb4502cfc74d76b0c6df2d9173855e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a38ce4af9b414c778a69cd3f64f5f6bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 736.575256] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662122, 'name': CreateVM_Task, 'duration_secs': 0.809735} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.575452] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 736.576025] env[65788]: WARNING neutronclient.v2_0.client [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 736.576535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.576887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.578074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 736.578074] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77067571-4bc7-4cfe-858b-518794583bc0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.584983] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 736.584983] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5275d1b5-6926-972f-b24a-4cab2ce7620a" [ 736.584983] env[65788]: _type = "Task" [ 736.584983] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.606458] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5275d1b5-6926-972f-b24a-4cab2ce7620a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.702383] env[65788]: DEBUG nova.network.neutron [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 736.770532] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662123, 'name': ReconfigVM_Task, 'duration_secs': 0.595307} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.773803] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Reconfigured VM instance instance-0000000a to attach disk [datastore2] volume-fe7e6d12-0fe6-42b3-847c-eb3860514545/volume-fe7e6d12-0fe6-42b3-847c-eb3860514545.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.779609] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-190640c6-7c89-4406-9882-08e06f5fa77b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.799896] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Waiting for the task: (returnval){ [ 736.799896] env[65788]: value = "task-4662125" [ 736.799896] env[65788]: _type = "Task" [ 736.799896] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.817582] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662125, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.826340] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662124, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.921379] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.921773] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 737.035028] env[65788]: DEBUG nova.network.neutron [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Successfully created port: 04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 737.102289] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5275d1b5-6926-972f-b24a-4cab2ce7620a, 'name': SearchDatastore_Task, 'duration_secs': 0.02769} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.104334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.104334] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 737.104334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.104334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.104517] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 737.104517] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37a15257-58c7-481c-8e14-77c116472da8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.112965] env[65788]: DEBUG nova.network.neutron [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 737.125929] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0125f01d-0d46-4a8e-a288-30bcb4ce917e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.130449] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 737.130792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 737.132552] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5913016-f9a7-4fba-97d7-321707f78529 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.140296] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb51669-5812-4dcb-8b8a-c726351e31b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.145864] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 737.145864] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a68d1f-6bc8-bdc3-f79f-27f9806e2bdf" [ 737.145864] env[65788]: _type = "Task" [ 737.145864] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.177573] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab2969e-4382-4fea-8104-e543c4efc401 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.184620] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a68d1f-6bc8-bdc3-f79f-27f9806e2bdf, 'name': SearchDatastore_Task, 'duration_secs': 0.01458} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.186445] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e323274b-31d9-48dc-b7f9-8e6935103eb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.192289] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1783422c-5970-47ee-ae03-7c8bc291cd84 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.197472] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 737.197472] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5236a004-94f2-2007-c113-e089aefdb5a3" [ 737.197472] env[65788]: _type = "Task" [ 737.197472] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.211118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Releasing lock "refresh_cache-3f20bd75-98ab-4798-834a-c0ffc3c6146d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.212437] env[65788]: DEBUG nova.compute.manager [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 737.212437] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.212437] env[65788]: DEBUG nova.compute.provider_tree [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.215026] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295ba613-1afd-46b5-af29-82abe0fb8edb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.222609] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5236a004-94f2-2007-c113-e089aefdb5a3, 'name': SearchDatastore_Task, 'duration_secs': 0.023547} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.224304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.224304] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 72111766-217d-469f-ad92-ca0ce0e67090/72111766-217d-469f-ad92-ca0ce0e67090.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 737.224908] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c54db3f-2ebc-44be-9d29-d34602ed8292 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.230152] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 737.230792] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4905b92b-60b9-4bdb-9a34-30e1828d592f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.236660] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 737.236660] env[65788]: value = "task-4662126" [ 737.236660] env[65788]: _type = "Task" [ 737.236660] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.241756] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 737.241756] env[65788]: value = "task-4662127" [ 737.241756] env[65788]: _type = "Task" [ 737.241756] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.250033] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662126, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.256428] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.311396] env[65788]: DEBUG oslo_vmware.api [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662125, 'name': ReconfigVM_Task, 'duration_secs': 0.218283} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.314555] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910232', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'name': 'volume-fe7e6d12-0fe6-42b3-847c-eb3860514545', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b701040-025c-4246-ad54-f2cf478e998d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'serial': 'fe7e6d12-0fe6-42b3-847c-eb3860514545'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 737.322580] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662124, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.471067] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 737.500252] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:47:39Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='edeb65e9-37d7-4c27-800f-7029f9e8ded0',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1773652449',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 737.500527] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 737.500682] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 737.500936] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 737.501112] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 737.501272] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 737.501481] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.501638] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 737.501860] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 737.501974] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 737.502185] env[65788]: DEBUG nova.virt.hardware [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 737.503230] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946cab7e-7db6-4b08-b1a6-5bb718f991d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.516059] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c8476b-45b0-48f7-98e9-26929b50b23c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.719854] env[65788]: DEBUG nova.scheduler.client.report [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 737.753777] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662126, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.761364] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662127, 'name': PowerOffVM_Task, 'duration_secs': 0.265123} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.761910] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 737.762235] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 737.762651] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef996382-afdc-4a10-a1d3-199ef1c5e461 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.780173] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 737.780173] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 737.797850] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 737.797850] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 737.797850] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Deleting the datastore file [datastore2] 3f20bd75-98ab-4798-834a-c0ffc3c6146d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 737.798340] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2ea42aa-f967-4977-85bb-4b9f2bd3bbd5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.808539] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for the task: (returnval){ [ 737.808539] env[65788]: value = "task-4662129" [ 737.808539] env[65788]: _type = "Task" [ 737.808539] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.830053] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.830053] env[65788]: DEBUG oslo_vmware.api [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662124, 'name': PowerOnVM_Task, 'duration_secs': 1.124238} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.830053] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 737.830053] env[65788]: DEBUG nova.compute.manager [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 737.830412] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e227cc4-89b8-460d-bde5-3fbcc8424674 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.226839] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.788s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.227429] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 738.230175] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.424s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 738.234404] env[65788]: INFO nova.compute.claims [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.246970] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662126, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670837} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.247346] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 72111766-217d-469f-ad92-ca0ce0e67090/72111766-217d-469f-ad92-ca0ce0e67090.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 738.247624] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 738.247925] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca4ca094-65bc-4126-b7a5-1bdeeedd4762 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.256977] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 738.256977] env[65788]: value = "task-4662130" [ 738.256977] env[65788]: _type = "Task" [ 738.256977] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.268640] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.322723] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.352574] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.417867] env[65788]: DEBUG nova.objects.instance [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lazy-loading 'flavor' on Instance uuid 5b701040-025c-4246-ad54-f2cf478e998d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.739698] env[65788]: DEBUG nova.network.neutron [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Successfully updated port: 04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 738.744649] env[65788]: DEBUG nova.compute.utils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 738.752596] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 738.752596] env[65788]: DEBUG nova.network.neutron [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 738.752596] env[65788]: WARNING neutronclient.v2_0.client [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.752596] env[65788]: WARNING neutronclient.v2_0.client [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.752596] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.752863] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.762890] env[65788]: WARNING neutronclient.v2_0.client [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.763088] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.763437] env[65788]: WARNING openstack [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.781817] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070259} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.782106] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.783422] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2168d4-0edf-42a9-b77a-794c95886ff0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.808372] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 72111766-217d-469f-ad92-ca0ce0e67090/72111766-217d-469f-ad92-ca0ce0e67090.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.808960] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a2c4099-e51b-4f7f-baf9-18ccdc8a5f1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.837526] env[65788]: DEBUG oslo_vmware.api [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Task: {'id': task-4662129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.673914} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.839745] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.839745] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 738.839745] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.839745] env[65788]: INFO nova.compute.manager [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Took 1.63 seconds to destroy the instance on the hypervisor. [ 738.840148] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 738.840364] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 738.840364] env[65788]: value = "task-4662131" [ 738.840364] env[65788]: _type = "Task" [ 738.840364] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.840616] env[65788]: DEBUG nova.compute.manager [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 738.840693] env[65788]: DEBUG nova.network.neutron [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 738.840922] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 738.841592] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.841854] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.862941] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662131, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.926413] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b1b63e88-c280-4ad7-9163-6cfcac3c9e6e tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lock "5b701040-025c-4246-ad54-f2cf478e998d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.909s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.981895] env[65788]: DEBUG nova.policy [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a1e3cda6bf9439982465aec80977cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1713c7002a413fb27180469fded83e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 739.114684] env[65788]: DEBUG nova.network.neutron [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 739.115133] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.223161] env[65788]: DEBUG nova.network.neutron [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Updating instance_info_cache with network_info: [{"id": "35fcd817-c75c-40df-beb8-b13fe638dd50", "address": "fa:16:3e:fd:7a:35", "network": {"id": "41adc3ec-0a47-4b0c-b506-4ade07c0b33e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1346818845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6002001018ca4f6983504dac34952e0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35fcd817-c7", "ovs_interfaceid": "35fcd817-c75c-40df-beb8-b13fe638dd50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 739.242225] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.244025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.244025] env[65788]: DEBUG nova.network.neutron [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 739.249672] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 739.365538] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662131, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.476974] env[65788]: DEBUG nova.network.neutron [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Successfully created port: a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 739.621626] env[65788]: DEBUG nova.network.neutron [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 739.729721] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Releasing lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 739.729810] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Instance network_info: |[{"id": "35fcd817-c75c-40df-beb8-b13fe638dd50", "address": "fa:16:3e:fd:7a:35", "network": {"id": "41adc3ec-0a47-4b0c-b506-4ade07c0b33e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1346818845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "6002001018ca4f6983504dac34952e0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35fcd817-c7", "ovs_interfaceid": "35fcd817-c75c-40df-beb8-b13fe638dd50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 739.730794] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:7a:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35fcd817-c75c-40df-beb8-b13fe638dd50', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 739.738675] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Creating folder: Project (6002001018ca4f6983504dac34952e0a). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 739.740786] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c00b796-e870-49ca-b48d-760eb8834a18 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.745536] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.745935] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.763794] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Created folder: Project (6002001018ca4f6983504dac34952e0a) in parent group-v910111. [ 739.763908] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Creating folder: Instances. Parent ref: group-v910237. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 739.768700] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb8cf4e7-855f-485b-9714-38e62940fad0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.783589] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Created folder: Instances in parent group-v910237. [ 739.783836] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 739.785078] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 739.785078] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd143137-9b24-443a-93de-50bd4322371e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.807617] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 739.807617] env[65788]: value = "task-4662134" [ 739.807617] env[65788]: _type = "Task" [ 739.807617] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.817094] env[65788]: DEBUG nova.network.neutron [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 739.821760] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662134, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.864781] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662131, 'name': ReconfigVM_Task, 'duration_secs': 0.63001} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.865950] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 72111766-217d-469f-ad92-ca0ce0e67090/72111766-217d-469f-ad92-ca0ce0e67090.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.865950] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a370bb50-aa6e-4cec-9e3a-3c2bd8e6c756 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.875278] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 739.875278] env[65788]: value = "task-4662135" [ 739.875278] env[65788]: _type = "Task" [ 739.875278] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.891389] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662135, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.903373] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.903768] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.914144] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ff3f39-9228-450b-99ef-0c97053cc53d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.925050] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e46274-afa7-497e-a9ed-0d7bda5875e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.965551] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35b0a38-adef-4b41-be42-e0685eb274a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.975276] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356e22e8-5e6c-40f7-9144-b0cc8f0cd370 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.994601] env[65788]: DEBUG nova.compute.provider_tree [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.030797] env[65788]: DEBUG nova.compute.manager [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Received event network-changed-3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 740.031191] env[65788]: DEBUG nova.compute.manager [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Refreshing instance network info cache due to event network-changed-3c6f2879-19ae-45b8-8785-45987a597a6a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 740.032128] env[65788]: DEBUG oslo_concurrency.lockutils [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Acquiring lock "refresh_cache-72111766-217d-469f-ad92-ca0ce0e67090" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.032128] env[65788]: DEBUG oslo_concurrency.lockutils [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Acquired lock "refresh_cache-72111766-217d-469f-ad92-ca0ce0e67090" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.032128] env[65788]: DEBUG nova.network.neutron [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Refreshing network info cache for port 3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 740.064180] env[65788]: WARNING neutronclient.v2_0.client [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.065033] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.065759] env[65788]: WARNING openstack [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.125385] env[65788]: INFO nova.compute.manager [-] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Took 1.28 seconds to deallocate network for instance. [ 740.240192] env[65788]: DEBUG nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Received event network-vif-plugged-35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 740.240260] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Acquiring lock "6be541bd-06fd-4959-9987-d250011b3f1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.240599] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Lock "6be541bd-06fd-4959-9987-d250011b3f1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 740.240836] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Lock "6be541bd-06fd-4959-9987-d250011b3f1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 740.241390] env[65788]: DEBUG nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] No waiting events found dispatching network-vif-plugged-35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 740.241647] env[65788]: WARNING nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Received unexpected event network-vif-plugged-35fcd817-c75c-40df-beb8-b13fe638dd50 for instance with vm_state building and task_state spawning. [ 740.241874] env[65788]: DEBUG nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Received event network-changed-35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 740.242267] env[65788]: DEBUG nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Refreshing instance network info cache due to event network-changed-35fcd817-c75c-40df-beb8-b13fe638dd50. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 740.242915] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Acquiring lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.244455] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Acquired lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.244455] env[65788]: DEBUG nova.network.neutron [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Refreshing network info cache for port 35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 740.271458] env[65788]: DEBUG nova.network.neutron [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.274447] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 740.309441] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 740.309884] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 740.310828] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 740.311167] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 740.311413] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 740.311621] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 740.311907] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.312154] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 740.312487] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 740.312705] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 740.312927] env[65788]: DEBUG nova.virt.hardware [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 740.314930] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b7ec2a-be5a-40ba-a557-0484ae886cc5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.337915] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5796c8d4-83f3-4d0e-9d2b-89572494a866 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.346235] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662134, 'name': CreateVM_Task, 'duration_secs': 0.39848} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.346563] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 740.347459] env[65788]: WARNING neutronclient.v2_0.client [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.347842] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.347998] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.348824] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 740.349172] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b4c1876-7a32-47c2-bc7e-5877eda597a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.365242] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 740.365242] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522f6736-85c2-3f0f-e4b0-f0cb6ba59d06" [ 740.365242] env[65788]: _type = "Task" [ 740.365242] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.374795] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522f6736-85c2-3f0f-e4b0-f0cb6ba59d06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.385995] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662135, 'name': Rename_Task, 'duration_secs': 0.185921} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.386319] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 740.386440] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6678f44e-e997-4e80-8446-06a510e8752f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.394656] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 740.394656] env[65788]: value = "task-4662136" [ 740.394656] env[65788]: _type = "Task" [ 740.394656] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.403180] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.500284] env[65788]: DEBUG nova.scheduler.client.report [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 740.537913] env[65788]: WARNING neutronclient.v2_0.client [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.540616] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.540616] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.633645] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 740.746544] env[65788]: WARNING neutronclient.v2_0.client [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 740.747238] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.747597] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.779598] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.779993] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Instance network_info: |[{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 740.780904] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:34:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04994181-be30-4c51-b4a1-ce198600a5c3', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.788910] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 740.789398] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.789650] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a53876c7-eaa2-4e7d-b5d2-faf6a0219b2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.811300] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.811300] env[65788]: value = "task-4662137" [ 740.811300] env[65788]: _type = "Task" [ 740.811300] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.820431] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662137, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.878758] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522f6736-85c2-3f0f-e4b0-f0cb6ba59d06, 'name': SearchDatastore_Task, 'duration_secs': 0.023201} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.879143] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.879405] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 740.879666] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.879858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.880070] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 740.880381] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afc28b71-cc4a-46db-bc93-6cb551636f96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.891376] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 740.892507] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 740.892507] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa58cf86-1400-4261-a1b2-9488572da69c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.902986] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 740.902986] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525f2e9c-4733-431a-c202-71f72dc61c7c" [ 740.902986] env[65788]: _type = "Task" [ 740.902986] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.907460] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662136, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.917979] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525f2e9c-4733-431a-c202-71f72dc61c7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.006251] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.776s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.006887] env[65788]: DEBUG nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 741.010547] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.975s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.010821] env[65788]: DEBUG nova.objects.instance [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 741.075277] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.075523] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.156410] env[65788]: DEBUG nova.network.neutron [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Successfully updated port: a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 741.268624] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.269119] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.327901] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662137, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.347187] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.347616] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.362549] env[65788]: WARNING neutronclient.v2_0.client [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.364987] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.364987] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.406077] env[65788]: DEBUG oslo_vmware.api [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662136, 'name': PowerOnVM_Task, 'duration_secs': 0.854438} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.406725] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 741.406983] env[65788]: INFO nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Took 9.03 seconds to spawn the instance on the hypervisor. [ 741.407230] env[65788]: DEBUG nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 741.408201] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e514d26-c4e7-4496-9948-3095bd8d05cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.429328] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525f2e9c-4733-431a-c202-71f72dc61c7c, 'name': SearchDatastore_Task, 'duration_secs': 0.01401} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.436021] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-201b5397-1dcd-490e-98c9-26f6012768a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.439880] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 741.439880] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525c8096-4543-cca5-dfaf-95b76b3504af" [ 741.439880] env[65788]: _type = "Task" [ 741.439880] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.444605] env[65788]: WARNING neutronclient.v2_0.client [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.445352] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.446098] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.460798] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525c8096-4543-cca5-dfaf-95b76b3504af, 'name': SearchDatastore_Task, 'duration_secs': 0.017108} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.461166] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.461606] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 6be541bd-06fd-4959-9987-d250011b3f1d/6be541bd-06fd-4959-9987-d250011b3f1d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 741.462304] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e3eade2-6f5e-4105-8fa2-9293cfb7c42f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.471625] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 741.471625] env[65788]: value = "task-4662138" [ 741.471625] env[65788]: _type = "Task" [ 741.471625] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.484841] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.512669] env[65788]: DEBUG nova.network.neutron [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Updated VIF entry in instance network info cache for port 3c6f2879-19ae-45b8-8785-45987a597a6a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 741.514356] env[65788]: DEBUG nova.network.neutron [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Updating instance_info_cache with network_info: [{"id": "3c6f2879-19ae-45b8-8785-45987a597a6a", "address": "fa:16:3e:3b:35:bb", "network": {"id": "934955b9-cb1d-49c3-9a97-76c56f073eab", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-107342658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8d2a28f70fdc4b07b9c4f3e9a6d39725", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3093647a-bab7-4562-ada0-428725e8c0fc", "external-id": "nsx-vlan-transportzone-660", "segmentation_id": 660, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c6f2879-19", "ovs_interfaceid": "3c6f2879-19ae-45b8-8785-45987a597a6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 741.517293] env[65788]: DEBUG nova.compute.utils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 741.522528] env[65788]: DEBUG nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 741.585417] env[65788]: DEBUG nova.network.neutron [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Updated VIF entry in instance network info cache for port 35fcd817-c75c-40df-beb8-b13fe638dd50. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 741.585869] env[65788]: DEBUG nova.network.neutron [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Updating instance_info_cache with network_info: [{"id": "35fcd817-c75c-40df-beb8-b13fe638dd50", "address": "fa:16:3e:fd:7a:35", "network": {"id": "41adc3ec-0a47-4b0c-b506-4ade07c0b33e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1346818845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6002001018ca4f6983504dac34952e0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35fcd817-c7", "ovs_interfaceid": "35fcd817-c75c-40df-beb8-b13fe638dd50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 741.662687] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "refresh_cache-99f2cc13-a559-4379-8332-d432e8bac36b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.662687] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "refresh_cache-99f2cc13-a559-4379-8332-d432e8bac36b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.662687] env[65788]: DEBUG nova.network.neutron [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 741.797246] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.797691] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.798073] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.798378] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.798631] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.802220] env[65788]: INFO nova.compute.manager [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Terminating instance [ 741.825725] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662137, 'name': CreateVM_Task, 'duration_secs': 0.566422} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.826063] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 741.827010] env[65788]: WARNING neutronclient.v2_0.client [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 741.827089] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.827335] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.827592] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 741.827853] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fc4033d-48fc-4e35-a361-d4252566b95f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.834780] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 741.834780] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d90f56-6d0b-1e87-ff58-15e8c1c63fda" [ 741.834780] env[65788]: _type = "Task" [ 741.834780] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.847827] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d90f56-6d0b-1e87-ff58-15e8c1c63fda, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.942099] env[65788]: INFO nova.compute.manager [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Took 33.88 seconds to build instance. [ 741.982866] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662138, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.016390] env[65788]: DEBUG oslo_concurrency.lockutils [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Releasing lock "refresh_cache-72111766-217d-469f-ad92-ca0ce0e67090" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.016694] env[65788]: DEBUG nova.compute.manager [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Received event network-changed-2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 742.016873] env[65788]: DEBUG nova.compute.manager [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Refreshing instance network info cache due to event network-changed-2117f600-f365-489d-9475-bf0afdc8f512. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 742.017246] env[65788]: DEBUG oslo_concurrency.lockutils [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Acquiring lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.017399] env[65788]: DEBUG oslo_concurrency.lockutils [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Acquired lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.017572] env[65788]: DEBUG nova.network.neutron [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Refreshing network info cache for port 2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 742.024758] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd51a561-af81-42e1-974c-e1b749e6c4e7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.026701] env[65788]: DEBUG nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 742.034024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.141s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.034024] env[65788]: DEBUG nova.objects.instance [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lazy-loading 'resources' on Instance uuid b5967423-89e1-44eb-b4de-22d09adb51a8 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 742.088752] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Releasing lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.089042] env[65788]: DEBUG nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-changed-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 742.089217] env[65788]: DEBUG nova.compute.manager [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing instance network info cache due to event network-changed-89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 742.089427] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Acquiring lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.089567] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Acquired lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.089730] env[65788]: DEBUG nova.network.neutron [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing network info cache for port 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 742.165649] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.166120] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.249238] env[65788]: DEBUG nova.network.neutron [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 742.273161] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.273814] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.310418] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "refresh_cache-0548399d-ec7a-4d0c-b436-9ef93c4e2d37" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.310620] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquired lock "refresh_cache-0548399d-ec7a-4d0c-b436-9ef93c4e2d37" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.310805] env[65788]: DEBUG nova.network.neutron [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 742.350451] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d90f56-6d0b-1e87-ff58-15e8c1c63fda, 'name': SearchDatastore_Task, 'duration_secs': 0.011517} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.350821] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.351156] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.351484] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.351684] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.351920] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.352284] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-939b6785-6e63-4db9-a00b-3b1c8b8bb1e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.359297] env[65788]: WARNING neutronclient.v2_0.client [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.359983] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.360399] env[65788]: WARNING openstack [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.369574] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.369762] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 742.370538] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b52957d4-5f4c-407a-90da-9652b9711a3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.379828] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 742.379828] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f85499-c01c-6b63-af84-e6a29c53e66b" [ 742.379828] env[65788]: _type = "Task" [ 742.379828] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.385892] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f85499-c01c-6b63-af84-e6a29c53e66b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.445127] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2437f957-ea2a-418e-99ee-21139b56d33e tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "72111766-217d-469f-ad92-ca0ce0e67090" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.922s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.481647] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.827336} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.482245] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 6be541bd-06fd-4959-9987-d250011b3f1d/6be541bd-06fd-4959-9987-d250011b3f1d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 742.482245] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 742.482508] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-749d0c55-d72b-48ab-92d6-e578cba080a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.486722] env[65788]: DEBUG nova.network.neutron [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Updating instance_info_cache with network_info: [{"id": "a490dde6-7a99-4085-9de3-e81116f08231", "address": "fa:16:3e:c5:fd:9d", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa490dde6-7a", "ovs_interfaceid": "a490dde6-7a99-4085-9de3-e81116f08231", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.494557] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 742.494557] env[65788]: value = "task-4662139" [ 742.494557] env[65788]: _type = "Task" [ 742.494557] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.511294] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662139, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.520771] env[65788]: WARNING neutronclient.v2_0.client [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.522068] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.522580] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.593284] env[65788]: WARNING neutronclient.v2_0.client [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.593284] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.593548] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.713971] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.714448] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.776157] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.776456] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.805819] env[65788]: WARNING neutronclient.v2_0.client [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.806624] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.806982] env[65788]: WARNING openstack [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.823458] env[65788]: WARNING neutronclient.v2_0.client [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.824144] env[65788]: WARNING openstack [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.824486] env[65788]: WARNING openstack [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.893479] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f85499-c01c-6b63-af84-e6a29c53e66b, 'name': SearchDatastore_Task, 'duration_secs': 0.021861} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.894507] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7eaf197e-99e5-477a-9046-289874210aa5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.905676] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 742.905676] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5279f0cf-9514-5e69-f9cb-4dbf6d70dfce" [ 742.905676] env[65788]: _type = "Task" [ 742.905676] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.925092] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5279f0cf-9514-5e69-f9cb-4dbf6d70dfce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.929079] env[65788]: DEBUG nova.network.neutron [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 742.949180] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 742.973560] env[65788]: WARNING neutronclient.v2_0.client [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.973560] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.973560] env[65788]: WARNING openstack [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.988893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "refresh_cache-99f2cc13-a559-4379-8332-d432e8bac36b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.989275] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Instance network_info: |[{"id": "a490dde6-7a99-4085-9de3-e81116f08231", "address": "fa:16:3e:c5:fd:9d", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa490dde6-7a", "ovs_interfaceid": "a490dde6-7a99-4085-9de3-e81116f08231", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 742.989741] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:fd:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a490dde6-7a99-4085-9de3-e81116f08231', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.000937] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 743.005252] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 743.010218] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7ff7f13-de33-439d-8c90-416c20ac079a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.041317] env[65788]: DEBUG nova.compute.manager [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Stashing vm_state: stopped {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 743.042811] env[65788]: DEBUG nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 743.051033] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662139, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071041} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.051387] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.051387] env[65788]: value = "task-4662140" [ 743.051387] env[65788]: _type = "Task" [ 743.051387] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.054277] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 743.055866] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8696cdd8-ee86-45dd-b732-6bb1459c6deb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.065031] env[65788]: DEBUG nova.network.neutron [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updated VIF entry in instance network info cache for port 2117f600-f365-489d-9475-bf0afdc8f512. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 743.065031] env[65788]: DEBUG nova.network.neutron [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updating instance_info_cache with network_info: [{"id": "2117f600-f365-489d-9475-bf0afdc8f512", "address": "fa:16:3e:1f:96:8c", "network": {"id": "8b4c23ed-2ee0-4fc1-b3bf-8306e97e5057", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1568181942-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a5e0d1546dd64e18a7ddf89faa3562dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2117f600-f3", "ovs_interfaceid": "2117f600-f365-489d-9475-bf0afdc8f512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.072315] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662140, 'name': CreateVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.097431] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 6be541bd-06fd-4959-9987-d250011b3f1d/6be541bd-06fd-4959-9987-d250011b3f1d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.108661] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b87fbc6d-3bd2-49d8-aed3-8cda8d84777b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.133289] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 743.133544] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 743.133688] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 743.133914] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 743.134080] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 743.134173] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 743.134398] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.134577] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 743.134743] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 743.134902] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 743.135110] env[65788]: DEBUG nova.virt.hardware [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 743.136309] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91939b2d-bb95-4d6c-94fa-67ff39027305 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.143597] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 743.143597] env[65788]: value = "task-4662141" [ 743.143597] env[65788]: _type = "Task" [ 743.143597] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.150959] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d166ff-5a80-4413-afc7-0193a59bd0dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.162818] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662141, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.175914] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "a813d102-1b77-4214-8eab-2cd66e002912" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.176209] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "a813d102-1b77-4214-8eab-2cd66e002912" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.176745] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.182432] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Creating folder: Project (4a9f792d6ee34bca8583bd4be402c4f9). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 743.186265] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a44f8001-df54-41eb-a4ce-3e911c756b28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.200178] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Created folder: Project (4a9f792d6ee34bca8583bd4be402c4f9) in parent group-v910111. [ 743.200379] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Creating folder: Instances. Parent ref: group-v910242. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 743.200630] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-842c6804-109b-47e6-8e82-bd05073a5351 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.223306] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Created folder: Instances in parent group-v910242. [ 743.223644] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 743.223898] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 743.224224] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd99d3d0-d33b-41e5-acc3-8380965cca58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.247514] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.247514] env[65788]: value = "task-4662144" [ 743.247514] env[65788]: _type = "Task" [ 743.247514] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.255169] env[65788]: DEBUG nova.network.neutron [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.267193] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662144, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.278671] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929c6567-bb78-47d1-b16d-c34deeb15fa0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.289438] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29f7ea3-db2e-46a4-8315-a7f16b96bd57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.296912] env[65788]: DEBUG nova.network.neutron [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updated VIF entry in instance network info cache for port 89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 743.297338] env[65788]: DEBUG nova.network.neutron [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 743.343631] env[65788]: DEBUG oslo_concurrency.lockutils [req-d829a933-641b-4b55-92e8-8e728cf5b709 req-3fb29d23-e807-4070-bd5c-c90a2bcf3656 service nova] Releasing lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.343631] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87873a4-d0c2-42d5-bc21-c3ea102876fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.353148] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6f9fc0-ecce-48dd-a2a8-a3fb3533a4dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.369551] env[65788]: DEBUG nova.compute.provider_tree [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.382225] env[65788]: DEBUG nova.compute.manager [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Received event network-vif-plugged-a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 743.382542] env[65788]: DEBUG oslo_concurrency.lockutils [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Acquiring lock "99f2cc13-a559-4379-8332-d432e8bac36b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.382612] env[65788]: DEBUG oslo_concurrency.lockutils [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Lock "99f2cc13-a559-4379-8332-d432e8bac36b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.382776] env[65788]: DEBUG oslo_concurrency.lockutils [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Lock "99f2cc13-a559-4379-8332-d432e8bac36b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.382932] env[65788]: DEBUG nova.compute.manager [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] No waiting events found dispatching network-vif-plugged-a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 743.383355] env[65788]: WARNING nova.compute.manager [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Received unexpected event network-vif-plugged-a490dde6-7a99-4085-9de3-e81116f08231 for instance with vm_state building and task_state spawning. [ 743.383355] env[65788]: DEBUG nova.compute.manager [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Received event network-changed-a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 743.383515] env[65788]: DEBUG nova.compute.manager [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Refreshing instance network info cache due to event network-changed-a490dde6-7a99-4085-9de3-e81116f08231. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 743.383747] env[65788]: DEBUG oslo_concurrency.lockutils [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Acquiring lock "refresh_cache-99f2cc13-a559-4379-8332-d432e8bac36b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.383747] env[65788]: DEBUG oslo_concurrency.lockutils [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Acquired lock "refresh_cache-99f2cc13-a559-4379-8332-d432e8bac36b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.384457] env[65788]: DEBUG nova.network.neutron [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Refreshing network info cache for port a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 743.414221] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Received event network-vif-plugged-04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 743.414321] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquiring lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.414530] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.414698] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.414858] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] No waiting events found dispatching network-vif-plugged-04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 743.415083] env[65788]: WARNING nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Received unexpected event network-vif-plugged-04994181-be30-4c51-b4a1-ce198600a5c3 for instance with vm_state building and task_state spawning. [ 743.415278] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Received event network-changed-04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 743.415468] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Refreshing instance network info cache due to event network-changed-04994181-be30-4c51-b4a1-ce198600a5c3. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 743.415755] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquiring lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.415814] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquired lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.415994] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Refreshing network info cache for port 04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 743.424290] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5279f0cf-9514-5e69-f9cb-4dbf6d70dfce, 'name': SearchDatastore_Task, 'duration_secs': 0.013923} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.425704] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.425704] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 743.425704] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4decdc68-c74f-425f-8561-6364abdf6446 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.433785] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 743.433785] env[65788]: value = "task-4662145" [ 743.433785] env[65788]: _type = "Task" [ 743.433785] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.445738] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.470377] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.569844] env[65788]: DEBUG oslo_concurrency.lockutils [req-19d546d9-06ec-4ed0-a9c8-c616d33ec428 req-a5f7284e-e942-4a41-9405-a75ad3260252 service nova] Releasing lock "refresh_cache-54ad3472-0cca-4dca-91f2-8c343f706926" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.569919] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662140, 'name': CreateVM_Task, 'duration_secs': 0.444265} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.571032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.571386] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 743.571952] env[65788]: WARNING neutronclient.v2_0.client [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.572349] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.572569] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.572794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 743.573078] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68fd83c4-5079-4d6d-9e12-2a777d0463f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.579332] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 743.579332] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c87b14-62ec-6bb9-2711-42e6860893c2" [ 743.579332] env[65788]: _type = "Task" [ 743.579332] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.588575] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c87b14-62ec-6bb9-2711-42e6860893c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.654945] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662141, 'name': ReconfigVM_Task, 'duration_secs': 0.442076} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.655252] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 6be541bd-06fd-4959-9987-d250011b3f1d/6be541bd-06fd-4959-9987-d250011b3f1d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.655893] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6d9a477-2dd5-4100-a65f-fdca2c84b813 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.663968] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 743.663968] env[65788]: value = "task-4662146" [ 743.663968] env[65788]: _type = "Task" [ 743.663968] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.675372] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662146, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.759759] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662144, 'name': CreateVM_Task, 'duration_secs': 0.351444} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.759950] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 743.760751] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.760751] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.760926] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 743.761548] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a54078fe-8e22-42d4-8047-15908cf8173e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.763137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Releasing lock "refresh_cache-0548399d-ec7a-4d0c-b436-9ef93c4e2d37" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.763544] env[65788]: DEBUG nova.compute.manager [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 743.763735] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.764610] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01799e51-40f3-433a-a978-c31c4868fb1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.772529] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 743.772529] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520a68b8-b6a7-b782-411d-fe628f8481b0" [ 743.772529] env[65788]: _type = "Task" [ 743.772529] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.775703] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 743.779275] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f937227-0a87-44b5-b2e1-2a0d2898e930 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.787813] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520a68b8-b6a7-b782-411d-fe628f8481b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.789448] env[65788]: DEBUG oslo_vmware.api [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 743.789448] env[65788]: value = "task-4662147" [ 743.789448] env[65788]: _type = "Task" [ 743.789448] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.798717] env[65788]: DEBUG oslo_vmware.api [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662147, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.873598] env[65788]: DEBUG nova.scheduler.client.report [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 743.889701] env[65788]: WARNING neutronclient.v2_0.client [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.891026] env[65788]: WARNING openstack [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.891026] env[65788]: WARNING openstack [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.919077] env[65788]: WARNING neutronclient.v2_0.client [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 743.919995] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.920460] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.947569] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662145, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.091975] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c87b14-62ec-6bb9-2711-42e6860893c2, 'name': SearchDatastore_Task, 'duration_secs': 0.01485} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.092741] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.093007] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.093303] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.093452] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.093636] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.093935] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd1e0939-f9a8-44d5-8a16-fce9e3bda41b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.113291] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.113291] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 744.113291] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2646d58-68ad-49de-a473-d96495f609c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.120106] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 744.120106] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a6d378-a567-3660-07fc-0236275f5035" [ 744.120106] env[65788]: _type = "Task" [ 744.120106] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.128894] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a6d378-a567-3660-07fc-0236275f5035, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.176414] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662146, 'name': Rename_Task, 'duration_secs': 0.156642} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.183397] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.183397] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8e02ea7-f10f-4ec8-88e5-49c63567ca8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.192126] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 744.192126] env[65788]: value = "task-4662148" [ 744.192126] env[65788]: _type = "Task" [ 744.192126] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.196862] env[65788]: WARNING openstack [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.197346] env[65788]: WARNING openstack [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.214469] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662148, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.258552] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.259106] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.298152] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520a68b8-b6a7-b782-411d-fe628f8481b0, 'name': SearchDatastore_Task, 'duration_secs': 0.017445} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.299109] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.299533] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.299910] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.300205] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.300522] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.300983] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3445ff49-290a-4d2e-801c-0d8dea980c69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.309191] env[65788]: DEBUG oslo_vmware.api [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662147, 'name': PowerOffVM_Task, 'duration_secs': 0.492148} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.309938] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 744.310146] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 744.310805] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e73368fa-d423-47f7-99d4-65f3a703ec25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.322034] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.322034] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 744.322034] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ad1491-69ec-4462-afaa-949f2a40218e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.328849] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 744.328849] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52125237-bb8b-8f11-5850-598ddcb70be5" [ 744.328849] env[65788]: _type = "Task" [ 744.328849] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.341465] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 744.341698] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 744.341881] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Deleting the datastore file [datastore2] 0548399d-ec7a-4d0c-b436-9ef93c4e2d37 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 744.342842] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-085efee3-c667-4606-adc2-ce7099a4294a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.350356] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52125237-bb8b-8f11-5850-598ddcb70be5, 'name': SearchDatastore_Task, 'duration_secs': 0.014735} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.352739] env[65788]: DEBUG oslo_vmware.api [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for the task: (returnval){ [ 744.352739] env[65788]: value = "task-4662150" [ 744.352739] env[65788]: _type = "Task" [ 744.352739] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.353093] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d557d276-626c-4b3d-b058-85209cd13840 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.364123] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 744.364123] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c65b7d-042b-740c-3b69-7f0f3bf83890" [ 744.364123] env[65788]: _type = "Task" [ 744.364123] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.370914] env[65788]: DEBUG oslo_vmware.api [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.372103] env[65788]: WARNING neutronclient.v2_0.client [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.372785] env[65788]: WARNING openstack [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.373176] env[65788]: WARNING openstack [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.382432] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.351s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.385113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.509s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.385376] env[65788]: DEBUG nova.objects.instance [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 744.392468] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c65b7d-042b-740c-3b69-7f0f3bf83890, 'name': SearchDatastore_Task, 'duration_secs': 0.01343} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.393044] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.393482] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 744.393778] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52e993da-613e-410b-b881-7bafeebed8ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.398851] env[65788]: WARNING neutronclient.v2_0.client [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.399624] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.400023] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.410747] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 744.410747] env[65788]: value = "task-4662151" [ 744.410747] env[65788]: _type = "Task" [ 744.410747] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.414679] env[65788]: INFO nova.scheduler.client.report [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Deleted allocations for instance b5967423-89e1-44eb-b4de-22d09adb51a8 [ 744.421923] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662151, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.450310] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.73768} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.450604] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 744.451166] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.451166] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7bdcc48e-8856-4e11-866c-ae6b96164eb3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.459069] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 744.459069] env[65788]: value = "task-4662152" [ 744.459069] env[65788]: _type = "Task" [ 744.459069] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.474365] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662152, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.525289] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.525693] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lock "5b701040-025c-4246-ad54-f2cf478e998d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.576568] env[65788]: DEBUG nova.network.neutron [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Updated VIF entry in instance network info cache for port a490dde6-7a99-4085-9de3-e81116f08231. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 744.577527] env[65788]: DEBUG nova.network.neutron [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Updating instance_info_cache with network_info: [{"id": "a490dde6-7a99-4085-9de3-e81116f08231", "address": "fa:16:3e:c5:fd:9d", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa490dde6-7a", "ovs_interfaceid": "a490dde6-7a99-4085-9de3-e81116f08231", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 744.591948] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "2c24bc30-d413-4714-81c2-b657cafe94bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.591948] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.634080] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a6d378-a567-3660-07fc-0236275f5035, 'name': SearchDatastore_Task, 'duration_secs': 0.03263} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.635254] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e799505c-dda4-4789-b6bf-9c267933cb1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.642495] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 744.642495] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527e189a-9c0e-720f-d80b-3e57ef94deaa" [ 744.642495] env[65788]: _type = "Task" [ 744.642495] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.644780] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updated VIF entry in instance network info cache for port 04994181-be30-4c51-b4a1-ce198600a5c3. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 744.645119] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 744.656685] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527e189a-9c0e-720f-d80b-3e57ef94deaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.702425] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662148, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.874645] env[65788]: DEBUG oslo_vmware.api [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Task: {'id': task-4662150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149886} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.875194] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 744.875194] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 744.875299] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.875516] env[65788]: INFO nova.compute.manager [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Took 1.11 seconds to destroy the instance on the hypervisor. [ 744.876021] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 744.876021] env[65788]: DEBUG nova.compute.manager [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 744.876021] env[65788]: DEBUG nova.network.neutron [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 744.876294] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.876813] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.877090] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.930785] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662151, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.931278] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e2ea6c8-d07e-45c0-b9d8-70dd6e47f44c tempest-ServerShowV247Test-1780817887 tempest-ServerShowV247Test-1780817887-project-member] Lock "b5967423-89e1-44eb-b4de-22d09adb51a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.322s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.965350] env[65788]: DEBUG nova.network.neutron [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 744.965615] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.969837] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662152, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112621} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.971990] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.973187] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1cbcf94-69da-446c-9509-3104c2211ca7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.999829] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.001071] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fa29b27-54c1-44d7-88b3-a89686fea0d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.023716] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 745.023716] env[65788]: value = "task-4662153" [ 745.023716] env[65788]: _type = "Task" [ 745.023716] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.029227] env[65788]: INFO nova.compute.manager [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Detaching volume fe7e6d12-0fe6-42b3-847c-eb3860514545 [ 745.038750] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662153, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.080355] env[65788]: DEBUG oslo_concurrency.lockutils [req-02adc2c6-2d41-403c-bb47-b2bead0cf57e req-757b3d00-5681-4120-8dfa-2485727b5b80 service nova] Releasing lock "refresh_cache-99f2cc13-a559-4379-8332-d432e8bac36b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.094883] env[65788]: INFO nova.virt.block_device [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Attempting to driver detach volume fe7e6d12-0fe6-42b3-847c-eb3860514545 from mountpoint /dev/sdb [ 745.095293] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 745.095400] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910232', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'name': 'volume-fe7e6d12-0fe6-42b3-847c-eb3860514545', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b701040-025c-4246-ad54-f2cf478e998d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'serial': 'fe7e6d12-0fe6-42b3-847c-eb3860514545'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 745.096299] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7824a9-9362-4910-9453-eeb531237669 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.124579] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf53357e-043c-4809-920a-fdee6523f794 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.132566] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f506006c-586d-40b7-8990-3de36e6f368f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.155267] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Releasing lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.155572] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-changed-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 745.155766] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing instance network info cache due to event network-changed-89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 745.155902] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquiring lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.156082] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquired lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.156224] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing network info cache for port 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 745.161388] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f562169b-2020-41af-8ead-1d2309f65bd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.174345] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527e189a-9c0e-720f-d80b-3e57ef94deaa, 'name': SearchDatastore_Task, 'duration_secs': 0.039671} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.187220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.187452] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 99f2cc13-a559-4379-8332-d432e8bac36b/99f2cc13-a559-4379-8332-d432e8bac36b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 745.188211] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] The volume has not been displaced from its original location: [datastore2] volume-fe7e6d12-0fe6-42b3-847c-eb3860514545/volume-fe7e6d12-0fe6-42b3-847c-eb3860514545.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 745.194228] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Reconfiguring VM instance instance-0000000a to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 745.194609] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e725998c-8b85-4b05-b62e-1cc948c642ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.197397] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffaa975d-ef25-4328-aeb3-050588061b83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.220921] env[65788]: DEBUG oslo_vmware.api [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662148, 'name': PowerOnVM_Task, 'duration_secs': 0.546116} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.222792] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.222792] env[65788]: INFO nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Took 10.55 seconds to spawn the instance on the hypervisor. [ 745.223022] env[65788]: DEBUG nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 745.223618] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Waiting for the task: (returnval){ [ 745.223618] env[65788]: value = "task-4662155" [ 745.223618] env[65788]: _type = "Task" [ 745.223618] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.225476] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447e8275-8ef7-4668-a628-56ba4dd1c369 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.228943] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 745.228943] env[65788]: value = "task-4662154" [ 745.228943] env[65788]: _type = "Task" [ 745.228943] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.249575] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662155, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.256021] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.400335] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c9b32da7-9e22-4264-9e96-c1257146ab32 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 745.401282] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.453s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.404951] env[65788]: INFO nova.compute.claims [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.421686] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662151, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.739737} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.422484] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 745.422484] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.422631] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c632a907-aa0a-4586-964d-8ca154a6ecc3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.431319] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 745.431319] env[65788]: value = "task-4662156" [ 745.431319] env[65788]: _type = "Task" [ 745.431319] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.441113] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662156, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.473689] env[65788]: DEBUG nova.network.neutron [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 745.536545] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662153, 'name': ReconfigVM_Task, 'duration_secs': 0.335625} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.536829] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfigured VM instance instance-00000028 to attach disk [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.540115] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-625df237-ee4f-4357-8009-31bc71de6497 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.545500] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 745.545500] env[65788]: value = "task-4662157" [ 745.545500] env[65788]: _type = "Task" [ 745.545500] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.557519] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662157, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.666446] env[65788]: WARNING neutronclient.v2_0.client [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.667328] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.667809] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.741916] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662155, 'name': ReconfigVM_Task, 'duration_secs': 0.511033} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.744971] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Reconfigured VM instance instance-0000000a to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 745.753806] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662154, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.758022] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da3e89a9-4c98-468f-ad7a-00dc34004553 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.767843] env[65788]: INFO nova.compute.manager [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Took 35.58 seconds to build instance. [ 745.781428] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Waiting for the task: (returnval){ [ 745.781428] env[65788]: value = "task-4662158" [ 745.781428] env[65788]: _type = "Task" [ 745.781428] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.793124] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662158, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.796678] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.796678] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.897034] env[65788]: WARNING neutronclient.v2_0.client [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 745.897536] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.897889] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.947513] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662156, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076089} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.947668] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.948542] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9da600-08ac-4bd8-acb9-aec793d896e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.979351] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Reconfiguring VM instance instance-0000002a to attach disk [datastore1] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.985206] env[65788]: INFO nova.compute.manager [-] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Took 1.11 seconds to deallocate network for instance. [ 745.985580] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c67789a-83c7-46cb-b7a1-4a5fb8ca614d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.012638] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 746.012638] env[65788]: value = "task-4662159" [ 746.012638] env[65788]: _type = "Task" [ 746.012638] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.027819] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662159, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.061949] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662157, 'name': Rename_Task, 'duration_secs': 0.161573} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.062243] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 746.062513] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9117b8d2-2e62-465f-931a-2221511a29d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.073482] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 746.073482] env[65788]: value = "task-4662160" [ 746.073482] env[65788]: _type = "Task" [ 746.073482] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.084257] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662160, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.101482] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updated VIF entry in instance network info cache for port 89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 746.101878] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.251672] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662154, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.010079} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.252086] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 99f2cc13-a559-4379-8332-d432e8bac36b/99f2cc13-a559-4379-8332-d432e8bac36b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 746.252376] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.252691] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-512f9cc9-3266-4be4-a970-2ab450ebb16b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.261163] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 746.261163] env[65788]: value = "task-4662161" [ 746.261163] env[65788]: _type = "Task" [ 746.261163] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.272192] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662161, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.274980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-398c87b9-72e8-4766-9962-7a9a33257b08 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.044s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.293622] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662158, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.505768] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.505768] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.514414] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 746.532340] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662159, 'name': ReconfigVM_Task, 'duration_secs': 0.42032} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.532340] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Reconfigured VM instance instance-0000002a to attach disk [datastore1] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.532776] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d68c9afe-395c-4a72-8f3d-32ea8289ce87 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.545835] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 746.545835] env[65788]: value = "task-4662162" [ 746.545835] env[65788]: _type = "Task" [ 746.545835] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.562933] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662162, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.573490] env[65788]: DEBUG nova.compute.manager [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Received event network-changed-325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 746.573490] env[65788]: DEBUG nova.compute.manager [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Refreshing instance network info cache due to event network-changed-325d6907-21b3-4ed1-9291-659b2ee9c13d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 746.573490] env[65788]: DEBUG oslo_concurrency.lockutils [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Acquiring lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.573490] env[65788]: DEBUG oslo_concurrency.lockutils [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Acquired lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 746.574032] env[65788]: DEBUG nova.network.neutron [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Refreshing network info cache for port 325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 746.596199] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662160, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.609411] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Releasing lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.609411] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Received event network-changed-325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 746.609633] env[65788]: DEBUG nova.compute.manager [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Refreshing instance network info cache due to event network-changed-325d6907-21b3-4ed1-9291-659b2ee9c13d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 746.610570] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquiring lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.779709] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 746.782607] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662161, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107757} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.789492] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 746.805021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52455975-39d5-494c-8508-86e5e31b3162 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.816157] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662158, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.841572] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 99f2cc13-a559-4379-8332-d432e8bac36b/99f2cc13-a559-4379-8332-d432e8bac36b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.847369] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-692e63f5-b084-40df-b7c8-61e213814ac0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.873841] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 746.873841] env[65788]: value = "task-4662163" [ 746.873841] env[65788]: _type = "Task" [ 746.873841] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.885328] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662163, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.062780] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662162, 'name': Rename_Task, 'duration_secs': 0.225295} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.063165] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 747.063395] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81d3b501-4c18-4a3a-8239-514085164e8b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.072792] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 747.072792] env[65788]: value = "task-4662164" [ 747.072792] env[65788]: _type = "Task" [ 747.072792] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.077866] env[65788]: WARNING neutronclient.v2_0.client [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.078563] env[65788]: WARNING openstack [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.078940] env[65788]: WARNING openstack [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.098024] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662164, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.098985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5418b5c-e779-421e-838d-ac666a8cfa13 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.113822] env[65788]: DEBUG oslo_vmware.api [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662160, 'name': PowerOnVM_Task, 'duration_secs': 0.554051} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.115929] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 747.116111] env[65788]: INFO nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Took 9.64 seconds to spawn the instance on the hypervisor. [ 747.116650] env[65788]: DEBUG nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 747.117141] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c45cc7-8053-4b2b-b171-9d92fffb7706 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.121458] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa548c13-1620-4aab-977b-e84406800b41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.166662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215db50c-e264-4fcf-b780-8ab4f233b122 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.180619] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13173826-5e2a-4410-ab0d-e1580a1d88ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.197573] env[65788]: DEBUG nova.compute.provider_tree [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.244026] env[65788]: WARNING openstack [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.244440] env[65788]: WARNING openstack [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.311561] env[65788]: DEBUG oslo_vmware.api [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Task: {'id': task-4662158, 'name': ReconfigVM_Task, 'duration_secs': 1.2232} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.311915] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910232', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'name': 'volume-fe7e6d12-0fe6-42b3-847c-eb3860514545', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '5b701040-025c-4246-ad54-f2cf478e998d', 'attached_at': '', 'detached_at': '', 'volume_id': 'fe7e6d12-0fe6-42b3-847c-eb3860514545', 'serial': 'fe7e6d12-0fe6-42b3-847c-eb3860514545'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 747.328271] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.343907] env[65788]: WARNING neutronclient.v2_0.client [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 747.344955] env[65788]: WARNING openstack [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.345651] env[65788]: WARNING openstack [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.386722] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662163, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.437689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "f604c16d-2a86-40d3-9891-5b33309b3047" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.438111] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.438268] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.438392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.438595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.441958] env[65788]: INFO nova.compute.manager [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Terminating instance [ 747.483074] env[65788]: DEBUG nova.network.neutron [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updated VIF entry in instance network info cache for port 325d6907-21b3-4ed1-9291-659b2ee9c13d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 747.483559] env[65788]: DEBUG nova.network.neutron [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [{"id": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "address": "fa:16:3e:ea:fa:21", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap325d6907-21", "ovs_interfaceid": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 747.586025] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662164, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.613566] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "caceb296-5da2-4b4d-b955-5d7a238ff939" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.614122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.674789] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "72111766-217d-469f-ad92-ca0ce0e67090" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.674995] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "72111766-217d-469f-ad92-ca0ce0e67090" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.675286] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "72111766-217d-469f-ad92-ca0ce0e67090-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.675506] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "72111766-217d-469f-ad92-ca0ce0e67090-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.675733] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "72111766-217d-469f-ad92-ca0ce0e67090-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.677591] env[65788]: INFO nova.compute.manager [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Took 36.77 seconds to build instance. [ 747.679123] env[65788]: INFO nova.compute.manager [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Terminating instance [ 747.701650] env[65788]: DEBUG nova.scheduler.client.report [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 747.896908] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662163, 'name': ReconfigVM_Task, 'duration_secs': 0.77147} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.897139] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 99f2cc13-a559-4379-8332-d432e8bac36b/99f2cc13-a559-4379-8332-d432e8bac36b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.898537] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a57f493-d4e0-431b-88e2-344fda4a2401 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.908832] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 747.908832] env[65788]: value = "task-4662165" [ 747.908832] env[65788]: _type = "Task" [ 747.908832] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.921956] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662165, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.947066] env[65788]: DEBUG nova.compute.manager [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 747.947658] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.948661] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a63560-6fef-4367-93e9-1ab61255f23a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.959791] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.960240] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69c047ca-56a8-4e78-bf47-0f1d4da19c8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.964431] env[65788]: DEBUG nova.objects.instance [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lazy-loading 'flavor' on Instance uuid 5b701040-025c-4246-ad54-f2cf478e998d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.974385] env[65788]: DEBUG oslo_vmware.api [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 747.974385] env[65788]: value = "task-4662166" [ 747.974385] env[65788]: _type = "Task" [ 747.974385] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.987510] env[65788]: DEBUG oslo_concurrency.lockutils [req-c7427357-6496-44b8-a2ac-0b8342452f3c req-68f85bb9-035b-4509-85ea-6bfaaafbc6ef service nova] Releasing lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.988032] env[65788]: DEBUG oslo_vmware.api [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662166, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.988299] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Acquired lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.988534] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Refreshing network info cache for port 325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 748.020115] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.020412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.084629] env[65788]: DEBUG oslo_vmware.api [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662164, 'name': PowerOnVM_Task, 'duration_secs': 0.784515} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.084828] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 748.085075] env[65788]: INFO nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Took 5.04 seconds to spawn the instance on the hypervisor. [ 748.085294] env[65788]: DEBUG nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 748.086142] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4750326-af03-4cc0-b8dc-fd7514fd859e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.179992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec9b5877-196d-4c7a-8259-609959abfcda tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.171s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.182411] env[65788]: DEBUG nova.compute.manager [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 748.182634] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.183733] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4ed9a1-6ab8-4f6a-ae83-ac81bd7ab575 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.196347] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 748.196347] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1adca056-f1dc-4f21-8825-3deca74825e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.203834] env[65788]: DEBUG oslo_vmware.api [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 748.203834] env[65788]: value = "task-4662167" [ 748.203834] env[65788]: _type = "Task" [ 748.203834] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.213072] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.812s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.213698] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 748.217370] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.429s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.219581] env[65788]: DEBUG nova.objects.instance [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'resources' on Instance uuid bcdf4a61-a2e8-4ca6-92fc-faddf5371290 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.227243] env[65788]: DEBUG oslo_vmware.api [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.420473] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662165, 'name': Rename_Task, 'duration_secs': 0.192181} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.420825] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 748.421162] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90002e55-7f49-4f6f-965c-5044f7506a48 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.429769] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 748.429769] env[65788]: value = "task-4662168" [ 748.429769] env[65788]: _type = "Task" [ 748.429769] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.439520] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662168, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.485034] env[65788]: DEBUG oslo_vmware.api [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662166, 'name': PowerOffVM_Task, 'duration_secs': 0.362901} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.485359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 748.485550] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 748.485822] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcdbab81-7318-44a7-bd60-5a0cc394a3b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.492583] env[65788]: WARNING neutronclient.v2_0.client [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.492986] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.493397] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.588973] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 748.589318] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 748.589399] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Deleting the datastore file [datastore2] f604c16d-2a86-40d3-9891-5b33309b3047 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.589686] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4d5b374-cc4b-480f-9a23-f2973994701f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.596513] env[65788]: DEBUG oslo_vmware.api [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 748.596513] env[65788]: value = "task-4662170" [ 748.596513] env[65788]: _type = "Task" [ 748.596513] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.601725] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.602387] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.624992] env[65788]: DEBUG oslo_vmware.api [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.624992] env[65788]: INFO nova.compute.manager [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Took 35.83 seconds to build instance. [ 748.679048] env[65788]: DEBUG oslo_concurrency.lockutils [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "6be541bd-06fd-4959-9987-d250011b3f1d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 748.679362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.679671] env[65788]: INFO nova.compute.manager [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Rebooting instance [ 748.683305] env[65788]: WARNING neutronclient.v2_0.client [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.683468] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.683704] env[65788]: WARNING openstack [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.693167] env[65788]: DEBUG nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 748.714861] env[65788]: DEBUG oslo_vmware.api [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662167, 'name': PowerOffVM_Task, 'duration_secs': 0.340621} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.715145] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 748.716393] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 748.716393] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70536d1c-7215-40d0-af1a-60975eeb45c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.722907] env[65788]: DEBUG nova.compute.utils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 748.729888] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 748.730056] env[65788]: DEBUG nova.network.neutron [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 748.730388] env[65788]: WARNING neutronclient.v2_0.client [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.730698] env[65788]: WARNING neutronclient.v2_0.client [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.731382] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.731746] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.789118] env[65788]: DEBUG nova.policy [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c028d9c053054a2f916507b237d37e60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06b3dfec7c0144928b4a59dbd27569b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 748.795031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 748.795304] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 748.795534] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Deleting the datastore file [datastore2] 72111766-217d-469f-ad92-ca0ce0e67090 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.798122] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updated VIF entry in instance network info cache for port 325d6907-21b3-4ed1-9291-659b2ee9c13d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 748.798806] env[65788]: DEBUG nova.network.neutron [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [{"id": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "address": "fa:16:3e:ea:fa:21", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap325d6907-21", "ovs_interfaceid": "325d6907-21b3-4ed1-9291-659b2ee9c13d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 748.799941] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f3d7a5d-7404-4d50-ad11-897a0b22ebd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.809283] env[65788]: DEBUG oslo_vmware.api [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for the task: (returnval){ [ 748.809283] env[65788]: value = "task-4662172" [ 748.809283] env[65788]: _type = "Task" [ 748.809283] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.821957] env[65788]: DEBUG oslo_vmware.api [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.946764] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662168, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.972790] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6b99cff3-713e-4436-ade5-a6c60c180f72 tempest-VolumesAssistedSnapshotsTest-1244566878 tempest-VolumesAssistedSnapshotsTest-1244566878-project-admin] Lock "5b701040-025c-4246-ad54-f2cf478e998d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.447s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.059846] env[65788]: DEBUG nova.compute.manager [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-changed-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 749.059846] env[65788]: DEBUG nova.compute.manager [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing instance network info cache due to event network-changed-89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 749.061074] env[65788]: DEBUG oslo_concurrency.lockutils [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Acquiring lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.061074] env[65788]: DEBUG oslo_concurrency.lockutils [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Acquired lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.061074] env[65788]: DEBUG nova.network.neutron [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing network info cache for port 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 749.111600] env[65788]: DEBUG oslo_vmware.api [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131447} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.111954] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 749.112180] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 749.112543] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.113280] env[65788]: INFO nova.compute.manager [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Took 1.17 seconds to destroy the instance on the hypervisor. [ 749.113280] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 749.113280] env[65788]: DEBUG nova.compute.manager [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 749.113488] env[65788]: DEBUG nova.network.neutron [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 749.113587] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.114251] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.114450] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.126545] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d95567f-1c9f-4229-9c6b-09c64836bb4d tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.189s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.219897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.223618] env[65788]: DEBUG oslo_concurrency.lockutils [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.223854] env[65788]: DEBUG oslo_concurrency.lockutils [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquired lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.224053] env[65788]: DEBUG nova.network.neutron [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 749.226857] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.231101] env[65788]: DEBUG nova.network.neutron [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Successfully created port: ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 749.234052] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 749.303550] env[65788]: DEBUG oslo_concurrency.lockutils [req-a041119b-ed5d-4259-a633-409752610d11 req-d696b567-e48c-4031-90d0-63a1ba6be988 service nova] Releasing lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 749.325950] env[65788]: DEBUG oslo_vmware.api [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Task: {'id': task-4662172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137851} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.326227] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 749.326996] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 749.326996] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.326996] env[65788]: INFO nova.compute.manager [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Took 1.14 seconds to destroy the instance on the hypervisor. [ 749.326996] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 749.327285] env[65788]: DEBUG nova.compute.manager [-] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 749.327285] env[65788]: DEBUG nova.network.neutron [-] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 749.327544] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.328408] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.328408] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.424287] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8b16d1-9e97-4dce-b388-fb7d322eda5f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.433180] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-734c515b-a6d0-47d6-9614-edfd97ad16ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.445323] env[65788]: DEBUG oslo_vmware.api [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662168, 'name': PowerOnVM_Task, 'duration_secs': 0.705472} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.482262] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.482403] env[65788]: INFO nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Took 9.21 seconds to spawn the instance on the hypervisor. [ 749.482594] env[65788]: DEBUG nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 749.485414] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17cb2d1a-c22e-40fb-bc91-c02b7f18032c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.488611] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34ed867-be20-4011-b4e3-a20d31482cc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.503572] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982969c9-cdbc-4862-afd7-82d2d0a48315 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.520706] env[65788]: DEBUG nova.compute.provider_tree [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.564118] env[65788]: WARNING neutronclient.v2_0.client [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.565148] env[65788]: WARNING openstack [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.566010] env[65788]: WARNING openstack [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.600452] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.630228] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 749.729684] env[65788]: WARNING neutronclient.v2_0.client [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 749.730531] env[65788]: WARNING openstack [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.731496] env[65788]: WARNING openstack [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.993048] env[65788]: DEBUG nova.network.neutron [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.013599] env[65788]: INFO nova.compute.manager [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Took 38.86 seconds to build instance. [ 750.019432] env[65788]: WARNING openstack [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.019604] env[65788]: WARNING openstack [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.031644] env[65788]: DEBUG nova.scheduler.client.report [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 750.040465] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f55f3ee1-5cf9-4da9-87d7-0c37d8162dbc tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "99f2cc13-a559-4379-8332-d432e8bac36b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.955s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.158342] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.220130] env[65788]: WARNING openstack [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.220613] env[65788]: WARNING openstack [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.246729] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 750.250657] env[65788]: WARNING neutronclient.v2_0.client [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.251456] env[65788]: WARNING openstack [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.251958] env[65788]: WARNING openstack [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.285941] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 750.286253] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 750.286365] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 750.286536] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 750.287012] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 750.287012] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 750.287109] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.287968] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 750.287968] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 750.287968] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 750.287968] env[65788]: DEBUG nova.virt.hardware [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 750.290771] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c25954-418d-49ce-ba2b-a155151a00ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.298206] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0067e746-3631-4505-99fe-b3925bbe6e9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.334472] env[65788]: DEBUG nova.compute.manager [req-4add6426-6dec-487a-899e-199fcfd8c796 req-7c876fe7-ba27-47b0-b5ea-81753dd1ace6 service nova] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Received event network-vif-deleted-325d6907-21b3-4ed1-9291-659b2ee9c13d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 750.336245] env[65788]: DEBUG nova.network.neutron [-] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.396459] env[65788]: WARNING neutronclient.v2_0.client [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 750.397334] env[65788]: WARNING openstack [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.397588] env[65788]: WARNING openstack [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.482405] env[65788]: DEBUG nova.network.neutron [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updated VIF entry in instance network info cache for port 89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 750.482936] env[65788]: DEBUG nova.network.neutron [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.494321] env[65788]: INFO nova.compute.manager [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Took 1.38 seconds to deallocate network for instance. [ 750.541835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.325s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.544961] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.886s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.546200] env[65788]: INFO nova.compute.claims [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.549580] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 750.578592] env[65788]: INFO nova.scheduler.client.report [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocations for instance bcdf4a61-a2e8-4ca6-92fc-faddf5371290 [ 750.590067] env[65788]: DEBUG nova.network.neutron [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Updating instance_info_cache with network_info: [{"id": "35fcd817-c75c-40df-beb8-b13fe638dd50", "address": "fa:16:3e:fd:7a:35", "network": {"id": "41adc3ec-0a47-4b0c-b506-4ade07c0b33e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1346818845-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6002001018ca4f6983504dac34952e0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35fcd817-c7", "ovs_interfaceid": "35fcd817-c75c-40df-beb8-b13fe638dd50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.660535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "77dcec96-50e5-4753-b3cb-c6aec377bce2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.660947] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.660947] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "77dcec96-50e5-4753-b3cb-c6aec377bce2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.661213] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.661402] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.663517] env[65788]: INFO nova.compute.manager [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Terminating instance [ 750.838956] env[65788]: INFO nova.compute.manager [-] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Took 1.51 seconds to deallocate network for instance. [ 750.879269] env[65788]: DEBUG nova.network.neutron [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Successfully updated port: ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 750.985923] env[65788]: DEBUG oslo_concurrency.lockutils [req-a9459bef-8e32-4310-bd43-d1b47deea5e2 req-60f3d841-bf1b-4d74-ac63-539502cd833d service nova] Releasing lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.001153] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.075868] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.087227] env[65788]: DEBUG oslo_concurrency.lockutils [None req-49d15ef4-3c5e-4da3-aa90-b22d3ec41eba tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "bcdf4a61-a2e8-4ca6-92fc-faddf5371290" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.800s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.093262] env[65788]: DEBUG oslo_concurrency.lockutils [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Releasing lock "refresh_cache-6be541bd-06fd-4959-9987-d250011b3f1d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.168087] env[65788]: DEBUG nova.compute.manager [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 751.168350] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.169492] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf6fc8f-43c4-4f7a-b256-00aec2c09503 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.181649] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 751.181966] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a27cc14-429a-44dd-8c39-80d06b156b7d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.190362] env[65788]: DEBUG oslo_vmware.api [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 751.190362] env[65788]: value = "task-4662174" [ 751.190362] env[65788]: _type = "Task" [ 751.190362] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.200585] env[65788]: DEBUG oslo_vmware.api [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662174, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.351303] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.382874] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.383049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.383264] env[65788]: DEBUG nova.network.neutron [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 751.463571] env[65788]: INFO nova.compute.manager [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Rebuilding instance [ 751.519021] env[65788]: DEBUG nova.compute.manager [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 751.519021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e166c23-bb8c-4987-9dbf-613a800e03fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.601191] env[65788]: DEBUG nova.compute.manager [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 751.602402] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedc7c96-1c75-4c9e-ae94-c010ed308c92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.705083] env[65788]: DEBUG oslo_vmware.api [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662174, 'name': PowerOffVM_Task, 'duration_secs': 0.316382} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.705083] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 751.705083] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 751.705083] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9e154d2-a85f-46ae-9b09-ac174c7e6c91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.781642] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 751.782093] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 751.782355] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Deleting the datastore file [datastore2] 77dcec96-50e5-4753-b3cb-c6aec377bce2 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.782767] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a444dcf4-d299-42cb-8577-0596c3982f9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.792836] env[65788]: DEBUG oslo_vmware.api [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 751.792836] env[65788]: value = "task-4662176" [ 751.792836] env[65788]: _type = "Task" [ 751.792836] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.809711] env[65788]: DEBUG oslo_vmware.api [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.889479] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.889479] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.988820] env[65788]: DEBUG nova.network.neutron [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 752.021823] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "bbf6a242-9873-44b9-8938-2f8df11a1018" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.022539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.024583] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.025223] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.111242] env[65788]: WARNING neutronclient.v2_0.client [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.111242] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.111242] env[65788]: WARNING openstack [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.171802] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcef41c-206f-473c-a6d1-069968c07c9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.179985] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Suspending the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 752.180292] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-0d08001a-291b-4f62-9248-296f0ba00921 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.185390] env[65788]: DEBUG nova.compute.manager [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Received event network-vif-plugged-ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 752.186027] env[65788]: DEBUG oslo_concurrency.lockutils [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Acquiring lock "f3c53f67-85bf-4c18-9313-75eb90862f78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.186682] env[65788]: DEBUG oslo_concurrency.lockutils [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.186682] env[65788]: DEBUG oslo_concurrency.lockutils [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.186682] env[65788]: DEBUG nova.compute.manager [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] No waiting events found dispatching network-vif-plugged-ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 752.186832] env[65788]: WARNING nova.compute.manager [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Received unexpected event network-vif-plugged-ea77fa15-d004-4f6d-bea0-02027383b034 for instance with vm_state building and task_state spawning. [ 752.187032] env[65788]: DEBUG nova.compute.manager [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Received event network-changed-ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 752.187292] env[65788]: DEBUG nova.compute.manager [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Refreshing instance network info cache due to event network-changed-ea77fa15-d004-4f6d-bea0-02027383b034. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 752.187615] env[65788]: DEBUG oslo_concurrency.lockutils [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Acquiring lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.192143] env[65788]: DEBUG oslo_vmware.api [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 752.192143] env[65788]: value = "task-4662177" [ 752.192143] env[65788]: _type = "Task" [ 752.192143] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.204848] env[65788]: DEBUG oslo_vmware.api [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662177, 'name': SuspendVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.216316] env[65788]: DEBUG nova.network.neutron [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Updating instance_info_cache with network_info: [{"id": "ea77fa15-d004-4f6d-bea0-02027383b034", "address": "fa:16:3e:eb:4d:86", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea77fa15-d0", "ovs_interfaceid": "ea77fa15-d004-4f6d-bea0-02027383b034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 752.262127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0dc3cff-1fd9-48af-bdf1-5295676fed09 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.272802] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859d8981-60f4-4f75-bcb0-0c50ed3d8c9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.312607] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c911a0a-bf53-4718-abd6-b9f5d785a90f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.325155] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eede8755-380d-445d-ba67-bbde6eb7e611 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.330818] env[65788]: DEBUG oslo_vmware.api [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145291} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.330818] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.330818] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 752.331217] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.331515] env[65788]: INFO nova.compute.manager [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Took 1.16 seconds to destroy the instance on the hypervisor. [ 752.331883] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 752.333468] env[65788]: DEBUG nova.compute.manager [-] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 752.333468] env[65788]: DEBUG nova.network.neutron [-] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 752.333634] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.334353] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.334688] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.350694] env[65788]: DEBUG nova.compute.provider_tree [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.377988] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 752.539987] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 752.540351] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d13e3137-8681-400d-a89e-20a131576cf4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.550471] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 752.550471] env[65788]: value = "task-4662178" [ 752.550471] env[65788]: _type = "Task" [ 752.550471] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.568134] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.628127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1021e430-c7f1-4bfc-abd9-19da32ad7c76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.637603] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Doing hard reboot of VM {{(pid=65788) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 752.637936] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-3d4d8fc0-0f7a-4442-91b9-279d2a94fdda {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.648058] env[65788]: DEBUG oslo_vmware.api [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 752.648058] env[65788]: value = "task-4662179" [ 752.648058] env[65788]: _type = "Task" [ 752.648058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.664059] env[65788]: DEBUG oslo_vmware.api [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662179, 'name': ResetVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.710698] env[65788]: DEBUG oslo_vmware.api [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662177, 'name': SuspendVM_Task} progress is 54%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.719629] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.720070] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Instance network_info: |[{"id": "ea77fa15-d004-4f6d-bea0-02027383b034", "address": "fa:16:3e:eb:4d:86", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea77fa15-d0", "ovs_interfaceid": "ea77fa15-d004-4f6d-bea0-02027383b034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 752.720867] env[65788]: DEBUG oslo_concurrency.lockutils [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Acquired lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.721113] env[65788]: DEBUG nova.network.neutron [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Refreshing network info cache for port ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 752.723060] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:4d:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ea77fa15-d004-4f6d-bea0-02027383b034', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.734580] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 752.735876] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.736345] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f70a083b-c573-4cc4-81cc-63981cf916c1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.764976] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.764976] env[65788]: value = "task-4662180" [ 752.764976] env[65788]: _type = "Task" [ 752.764976] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.780371] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662180, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.858143] env[65788]: DEBUG nova.scheduler.client.report [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 753.067324] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662178, 'name': PowerOffVM_Task, 'duration_secs': 0.177396} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.067642] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 753.067877] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.068780] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc884cd-bee5-4151-aeb6-c42d2a549d36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.081456] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 753.085762] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16c48edb-e004-462f-9954-68c51e7ce577 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.121809] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 753.122229] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 753.122512] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Deleting the datastore file [datastore1] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 753.124704] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96fc8d8b-fedb-4c7f-8dda-24f867b22d12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.138612] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 753.138612] env[65788]: value = "task-4662182" [ 753.138612] env[65788]: _type = "Task" [ 753.138612] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.146675] env[65788]: DEBUG nova.compute.manager [req-bd56d616-ba1a-49d3-86fc-7cd2775f7da4 req-09c6b237-dcc5-45c2-a556-7adcfdccf114 service nova] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Received event network-vif-deleted-3c6f2879-19ae-45b8-8785-45987a597a6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 753.157929] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.165939] env[65788]: DEBUG oslo_vmware.api [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662179, 'name': ResetVM_Task, 'duration_secs': 0.121074} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.166262] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Did hard reboot of VM {{(pid=65788) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 753.166419] env[65788]: DEBUG nova.compute.manager [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 753.167272] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1247fbb9-56df-4931-9b8d-cef9352d79dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.210423] env[65788]: DEBUG oslo_vmware.api [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662177, 'name': SuspendVM_Task, 'duration_secs': 0.932502} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.210824] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Suspended the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 753.212081] env[65788]: DEBUG nova.compute.manager [None req-cb3c3936-e187-485f-a834-be0e78f66753 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 753.212081] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059bfc48-bc0b-4770-8fdd-49674d43ae42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.229191] env[65788]: WARNING neutronclient.v2_0.client [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.232774] env[65788]: WARNING openstack [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.232774] env[65788]: WARNING openstack [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.279916] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662180, 'name': CreateVM_Task, 'duration_secs': 0.458563} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.280343] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 753.282335] env[65788]: WARNING neutronclient.v2_0.client [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.282443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.282540] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.282897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 753.283270] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2eefd165-236a-4b18-85a3-f3009501e2f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.289488] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 753.289488] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e6f2d0-dc72-e9d2-15c7-0b8976d6b7e9" [ 753.289488] env[65788]: _type = "Task" [ 753.289488] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.306051] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e6f2d0-dc72-e9d2-15c7-0b8976d6b7e9, 'name': SearchDatastore_Task, 'duration_secs': 0.013378} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.306230] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.306561] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.306817] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.306967] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.307187] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.310258] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34e19520-6b3d-4a90-88ae-98dde89b4ec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.320394] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.320585] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.321946] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f0802c8-3aa5-47da-8668-6a6edcb7c132 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.330128] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 753.330128] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5293b7db-8323-36b8-ff94-635525b7384d" [ 753.330128] env[65788]: _type = "Task" [ 753.330128] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.339588] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5293b7db-8323-36b8-ff94-635525b7384d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.366314] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.822s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.367128] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 753.370612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.581s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.372686] env[65788]: INFO nova.compute.claims [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.386037] env[65788]: WARNING openstack [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.386726] env[65788]: WARNING openstack [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.453976] env[65788]: WARNING neutronclient.v2_0.client [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.454750] env[65788]: WARNING openstack [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.455023] env[65788]: WARNING openstack [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.557247] env[65788]: DEBUG nova.network.neutron [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Updated VIF entry in instance network info cache for port ea77fa15-d004-4f6d-bea0-02027383b034. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 753.557831] env[65788]: DEBUG nova.network.neutron [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Updating instance_info_cache with network_info: [{"id": "ea77fa15-d004-4f6d-bea0-02027383b034", "address": "fa:16:3e:eb:4d:86", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea77fa15-d0", "ovs_interfaceid": "ea77fa15-d004-4f6d-bea0-02027383b034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.595211] env[65788]: DEBUG nova.network.neutron [-] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.606599] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.606878] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "5b701040-025c-4246-ad54-f2cf478e998d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.608944] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "5b701040-025c-4246-ad54-f2cf478e998d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.608944] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "5b701040-025c-4246-ad54-f2cf478e998d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.608944] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "5b701040-025c-4246-ad54-f2cf478e998d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.610627] env[65788]: INFO nova.compute.manager [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Terminating instance [ 753.632131] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "f604c16d-2a86-40d3-9891-5b33309b3047" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.651460] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168993} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.653489] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.653726] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 753.653927] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 753.683982] env[65788]: DEBUG oslo_concurrency.lockutils [None req-97e253b6-bc28-4e5f-b1eb-e837775bd59f tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.005s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.842256] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5293b7db-8323-36b8-ff94-635525b7384d, 'name': SearchDatastore_Task, 'duration_secs': 0.022943} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.843105] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d35e34d1-1504-4dc7-bda6-9984343fcbe9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.849793] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 753.849793] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a0fd78-29a5-f4f9-8b0a-fa3d12f684cc" [ 753.849793] env[65788]: _type = "Task" [ 753.849793] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.858545] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a0fd78-29a5-f4f9-8b0a-fa3d12f684cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.879775] env[65788]: DEBUG nova.compute.utils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 753.884484] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 753.885098] env[65788]: DEBUG nova.network.neutron [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 753.885098] env[65788]: WARNING neutronclient.v2_0.client [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.885393] env[65788]: WARNING neutronclient.v2_0.client [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.886800] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.886800] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.951160] env[65788]: DEBUG nova.policy [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9310d3b6bf5430289df41fc742584c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4cffe8dfc424459a90e316ac4ac905f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 754.018024] env[65788]: DEBUG nova.compute.manager [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 754.060755] env[65788]: DEBUG oslo_concurrency.lockutils [req-c3191744-17e4-4df0-985e-cb062dbe8774 req-59685fb3-599d-4ea9-a846-692d7eb1b01c service nova] Releasing lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.097408] env[65788]: INFO nova.compute.manager [-] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Took 1.76 seconds to deallocate network for instance. [ 754.116047] env[65788]: DEBUG nova.compute.manager [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 754.116303] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 754.117237] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047f3d75-ce57-4624-88d9-d69f7c80df8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.126675] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 754.127251] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7dce365-8518-4d13-b0e4-9a08634b383b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.134871] env[65788]: DEBUG oslo_vmware.api [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 754.134871] env[65788]: value = "task-4662183" [ 754.134871] env[65788]: _type = "Task" [ 754.134871] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.151257] env[65788]: DEBUG oslo_vmware.api [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4662183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.336032] env[65788]: DEBUG nova.network.neutron [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Successfully created port: f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 754.362299] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a0fd78-29a5-f4f9-8b0a-fa3d12f684cc, 'name': SearchDatastore_Task, 'duration_secs': 0.009795} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.362819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.363317] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f3c53f67-85bf-4c18-9313-75eb90862f78/f3c53f67-85bf-4c18-9313-75eb90862f78.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 754.363630] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc9f0c6d-d70d-415b-8190-8cd9d005b2d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.374295] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 754.374295] env[65788]: value = "task-4662184" [ 754.374295] env[65788]: _type = "Task" [ 754.374295] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.387394] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 754.391470] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.539331] env[65788]: DEBUG nova.compute.manager [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-changed-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 754.541458] env[65788]: DEBUG nova.compute.manager [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing instance network info cache due to event network-changed-89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 754.541458] env[65788]: DEBUG oslo_concurrency.lockutils [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Acquiring lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.541698] env[65788]: DEBUG oslo_concurrency.lockutils [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Acquired lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.541753] env[65788]: DEBUG nova.network.neutron [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Refreshing network info cache for port 89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 754.553493] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.605902] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.647058] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "6aa4b17b-816c-4d84-8f74-a81185f3af65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.647373] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 754.647580] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "6aa4b17b-816c-4d84-8f74-a81185f3af65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 754.647796] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 754.647993] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.650342] env[65788]: DEBUG oslo_vmware.api [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4662183, 'name': PowerOffVM_Task, 'duration_secs': 0.3192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.654560] env[65788]: INFO nova.compute.manager [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Terminating instance [ 754.655484] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 754.655782] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 754.659461] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-adcd40a7-5f88-476a-9429-0771cdcb33ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.698585] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 754.699476] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 754.699634] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 754.699864] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 754.700046] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 754.700208] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 754.700567] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.700909] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 754.701401] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 754.701844] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 754.702054] env[65788]: DEBUG nova.virt.hardware [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 754.703912] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d5e80b-0e0d-495b-bf50-82ff362cb754 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.720435] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227d24f1-0f69-4125-9959-c87ca081bdba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.739645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.746017] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.751441] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.751788] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 754.751989] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 754.752260] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Deleting the datastore file [datastore1] 5b701040-025c-4246-ad54-f2cf478e998d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.753995] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3985fab-69fe-43e3-835e-bb13ad32df8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.766440] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37072d54-c238-4e1d-a692-c515533f13c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.777118] env[65788]: DEBUG oslo_vmware.api [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for the task: (returnval){ [ 754.777118] env[65788]: value = "task-4662186" [ 754.777118] env[65788]: _type = "Task" [ 754.777118] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.778266] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.778266] env[65788]: value = "task-4662187" [ 754.778266] env[65788]: _type = "Task" [ 754.778266] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.800605] env[65788]: DEBUG oslo_vmware.api [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4662186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.800819] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662187, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.890293] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662184, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.047165] env[65788]: WARNING neutronclient.v2_0.client [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.048353] env[65788]: WARNING openstack [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.049057] env[65788]: WARNING openstack [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.132453] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a49243b-3101-404f-b0a0-095ac81477d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.149112] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dcb908d-4be5-48f5-8471-5efe6a039feb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.195532] env[65788]: DEBUG nova.compute.manager [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 755.195743] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.201339] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b628a36-1d41-467f-860b-848c7908699b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.204573] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0210b2fe-8a88-47eb-a5eb-540f3a8251a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.218678] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80259983-dc71-4f67-b1e5-d8b57481c97a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.222851] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.224203] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd63cb56-f237-4197-b3b1-e9fc0df0ec14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.230192] env[65788]: WARNING openstack [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.230584] env[65788]: WARNING openstack [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.246806] env[65788]: DEBUG nova.compute.provider_tree [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.251119] env[65788]: DEBUG oslo_vmware.api [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 755.251119] env[65788]: value = "task-4662188" [ 755.251119] env[65788]: _type = "Task" [ 755.251119] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.265802] env[65788]: DEBUG oslo_vmware.api [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.299694] env[65788]: DEBUG oslo_vmware.api [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Task: {'id': task-4662186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.441709} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.299917] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662187, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.300185] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 755.300366] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 755.300541] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.300711] env[65788]: INFO nova.compute.manager [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Took 1.18 seconds to destroy the instance on the hypervisor. [ 755.300982] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 755.301332] env[65788]: DEBUG nova.compute.manager [-] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 755.301435] env[65788]: DEBUG nova.network.neutron [-] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 755.301882] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.302497] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.302804] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.386705] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60776} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.387069] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f3c53f67-85bf-4c18-9313-75eb90862f78/f3c53f67-85bf-4c18-9313-75eb90862f78.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 755.387337] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.388780] env[65788]: WARNING neutronclient.v2_0.client [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.389530] env[65788]: WARNING openstack [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.389941] env[65788]: WARNING openstack [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.398687] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-285016be-fcc3-42b9-b23b-dff3ea954d2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.402702] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 755.412731] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 755.412731] env[65788]: value = "task-4662189" [ 755.412731] env[65788]: _type = "Task" [ 755.412731] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.425357] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662189, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.433114] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 755.439153] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 755.439153] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.439153] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 755.439324] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.439324] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 755.439447] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 755.439812] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.440124] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 755.440426] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 755.440704] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 755.441037] env[65788]: DEBUG nova.virt.hardware [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 755.442111] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0536057d-d837-4b75-b4a8-4074f5fa523b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.451793] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb906a3f-2328-49bb-9123-df9fdba5ce89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.589346] env[65788]: DEBUG nova.compute.manager [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 755.590169] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfbf041a-2027-47aa-b369-e58198dd371f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.729811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "6be541bd-06fd-4959-9987-d250011b3f1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.730101] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.730311] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "6be541bd-06fd-4959-9987-d250011b3f1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.730485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.730643] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.733040] env[65788]: INFO nova.compute.manager [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Terminating instance [ 755.753560] env[65788]: DEBUG nova.scheduler.client.report [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 755.767435] env[65788]: DEBUG oslo_vmware.api [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662188, 'name': PowerOffVM_Task, 'duration_secs': 0.461059} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.767857] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.768031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 755.768305] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d7cba1a-7eac-41ca-8220-f9ed0c56b474 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.797480] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662187, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.922848] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662189, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075868} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.923506] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.924148] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae8a83b-7dda-4e62-b58c-8c2355831b02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.949942] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] f3c53f67-85bf-4c18-9313-75eb90862f78/f3c53f67-85bf-4c18-9313-75eb90862f78.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.951593] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b192b6c3-0ebd-476c-abcd-73b7e5df18c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.966371] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 755.966613] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 755.966933] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Deleting the datastore file [datastore1] 6aa4b17b-816c-4d84-8f74-a81185f3af65 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 755.967382] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f44e8e13-0ede-4134-bb2e-10f8cff69424 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.976147] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 755.976147] env[65788]: value = "task-4662191" [ 755.976147] env[65788]: _type = "Task" [ 755.976147] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.977623] env[65788]: DEBUG oslo_vmware.api [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for the task: (returnval){ [ 755.977623] env[65788]: value = "task-4662192" [ 755.977623] env[65788]: _type = "Task" [ 755.977623] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.990743] env[65788]: DEBUG oslo_vmware.api [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.994388] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662191, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.013732] env[65788]: DEBUG nova.network.neutron [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Successfully updated port: f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 756.067622] env[65788]: DEBUG nova.network.neutron [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updated VIF entry in instance network info cache for port 89c8bddf-8224-4440-87da-352af62539fc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 756.068314] env[65788]: DEBUG nova.network.neutron [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [{"id": "89c8bddf-8224-4440-87da-352af62539fc", "address": "fa:16:3e:23:3a:04", "network": {"id": "fa904338-a67c-43e9-877c-d734615ff65e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1517248639-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89909e0b9fd9487982d43248f9022481", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "136c3499-9ca0-4f85-903d-1f194aa66ed9", "external-id": "nsx-vlan-transportzone-307", "segmentation_id": 307, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89c8bddf-82", "ovs_interfaceid": "89c8bddf-8224-4440-87da-352af62539fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.105019] env[65788]: INFO nova.compute.manager [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] instance snapshotting [ 756.105019] env[65788]: WARNING nova.compute.manager [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 756.105448] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd529d1-2a6f-477a-a876-6a1df45ce34e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.132986] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dec73a-124b-41da-a501-b6e67c7e7fc8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.236913] env[65788]: DEBUG nova.compute.manager [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 756.237181] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 756.238128] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c01701-0615-474d-a433-9c53430d9dbf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.247650] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 756.247795] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8ec1b9b-1d8d-4aa0-8904-a6dabcd919e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.258029] env[65788]: DEBUG oslo_vmware.api [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 756.258029] env[65788]: value = "task-4662193" [ 756.258029] env[65788]: _type = "Task" [ 756.258029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.258029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.887s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.258029] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 756.266163] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.374s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.267909] env[65788]: INFO nova.compute.claims [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.278902] env[65788]: DEBUG oslo_vmware.api [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.049647] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.049954] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.050045] env[65788]: DEBUG nova.network.neutron [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 757.051223] env[65788]: DEBUG nova.network.neutron [-] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.052352] env[65788]: DEBUG oslo_concurrency.lockutils [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] Releasing lock "refresh_cache-12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.052584] env[65788]: DEBUG nova.compute.manager [req-a2da65bb-ff6c-496a-964e-9513e371c539 req-cae35f8c-0be5-4fc4-a1bb-4e2dd6a2236d service nova] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Received event network-vif-deleted-ca17fd35-e868-4f50-ac40-b6da3a12b015 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 757.053682] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 757.054842] env[65788]: DEBUG nova.compute.utils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 757.058501] env[65788]: DEBUG nova.compute.manager [req-9c6ea331-7176-4ff4-bec8-9b594585feac req-d23637f7-1433-4849-8fd8-bb5ef15cce27 service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Received event network-vif-deleted-709378e7-1b93-4d0c-9649-3fb90a38b091 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 757.058581] env[65788]: INFO nova.compute.manager [req-9c6ea331-7176-4ff4-bec8-9b594585feac req-d23637f7-1433-4849-8fd8-bb5ef15cce27 service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Neutron deleted interface 709378e7-1b93-4d0c-9649-3fb90a38b091; detaching it from the instance and deleting it from the info cache [ 757.058693] env[65788]: DEBUG nova.network.neutron [req-9c6ea331-7176-4ff4-bec8-9b594585feac req-d23637f7-1433-4849-8fd8-bb5ef15cce27 service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.060784] env[65788]: DEBUG nova.compute.manager [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Received event network-vif-plugged-f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 757.060784] env[65788]: DEBUG oslo_concurrency.lockutils [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Acquiring lock "9438ab56-1b4c-4778-a608-de319ab0ee43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.060908] env[65788]: DEBUG oslo_concurrency.lockutils [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.061055] env[65788]: DEBUG oslo_concurrency.lockutils [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 757.061190] env[65788]: DEBUG nova.compute.manager [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] No waiting events found dispatching network-vif-plugged-f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 757.061353] env[65788]: WARNING nova.compute.manager [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Received unexpected event network-vif-plugged-f116c9d8-4c98-4af9-8747-0c585d99d738 for instance with vm_state building and task_state spawning. [ 757.061505] env[65788]: DEBUG nova.compute.manager [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Received event network-changed-f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 757.061651] env[65788]: DEBUG nova.compute.manager [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Refreshing instance network info cache due to event network-changed-f116c9d8-4c98-4af9-8747-0c585d99d738. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 757.061811] env[65788]: DEBUG oslo_concurrency.lockutils [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Acquiring lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.068867] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2fe7bd4d-45d0-4955-8dd5-e6a4cc511b00 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.074081] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 757.074209] env[65788]: DEBUG nova.network.neutron [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 757.074521] env[65788]: WARNING neutronclient.v2_0.client [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.074832] env[65788]: WARNING neutronclient.v2_0.client [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.075405] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.075745] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.095426] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662187, 'name': CreateVM_Task, 'duration_secs': 1.416816} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.105206] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.107417] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662191, 'name': ReconfigVM_Task, 'duration_secs': 0.299707} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.107417] env[65788]: DEBUG oslo_vmware.api [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Task: {'id': task-4662192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158826} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.107417] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 757.107417] env[65788]: value = "task-4662194" [ 757.107417] env[65788]: _type = "Task" [ 757.107417] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.110587] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.110751] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.111094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 757.111381] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Reconfigured VM instance instance-0000002b to attach disk [datastore2] f3c53f67-85bf-4c18-9313-75eb90862f78/f3c53f67-85bf-4c18-9313-75eb90862f78.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.111995] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 757.112188] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 757.112377] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.112539] env[65788]: INFO nova.compute.manager [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Took 1.92 seconds to destroy the instance on the hypervisor. [ 757.112768] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 757.113613] env[65788]: DEBUG oslo_vmware.api [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662193, 'name': PowerOffVM_Task, 'duration_secs': 0.579598} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.113839] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec7060e2-dbf5-4734-91b4-0ddbac299b12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.115776] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-089dea37-743b-4197-a0d8-b26ddcc27fa5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.117600] env[65788]: DEBUG nova.compute.manager [-] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 757.117717] env[65788]: DEBUG nova.network.neutron [-] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 757.118033] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.119187] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.119456] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.129686] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 757.129963] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 757.135932] env[65788]: DEBUG nova.policy [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '49bb4c3a0a774c16a54d48847f56e6af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81015cf240554f4b899ba703842162bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 757.136914] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76f1f81e-2773-40e8-baaa-9b7f03737f80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.143125] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 757.143125] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5272e93b-3e02-87e8-10a9-e902d2b997d4" [ 757.143125] env[65788]: _type = "Task" [ 757.143125] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.149884] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662194, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.154416] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 757.154416] env[65788]: value = "task-4662195" [ 757.154416] env[65788]: _type = "Task" [ 757.154416] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.162528] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5272e93b-3e02-87e8-10a9-e902d2b997d4, 'name': SearchDatastore_Task, 'duration_secs': 0.013527} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.163239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.163475] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.163719] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.163864] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.164068] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.164338] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db3b7c7f-a9cd-4cd3-adf6-5219910515fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.171924] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662195, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.173262] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.183561] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.183764] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.184765] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2354c0ca-ec3a-4bb4-87fc-4e7d12677700 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.192588] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 757.192588] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522a1bc9-4a1e-2528-aac8-0eabb972e02f" [ 757.192588] env[65788]: _type = "Task" [ 757.192588] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.207513] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522a1bc9-4a1e-2528-aac8-0eabb972e02f, 'name': SearchDatastore_Task, 'duration_secs': 0.011631} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.208325] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c59aa2a8-af00-4d7b-82f2-8c94c9d47571 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.212127] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 757.212372] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 757.212533] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Deleting the datastore file [datastore2] 6be541bd-06fd-4959-9987-d250011b3f1d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 757.213174] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82bff370-11a4-4346-a1da-bc0dca946365 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.216142] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 757.216142] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52de3e56-757f-ab76-6544-8b67147ee250" [ 757.216142] env[65788]: _type = "Task" [ 757.216142] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.221282] env[65788]: DEBUG oslo_vmware.api [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for the task: (returnval){ [ 757.221282] env[65788]: value = "task-4662197" [ 757.221282] env[65788]: _type = "Task" [ 757.221282] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.227395] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52de3e56-757f-ab76-6544-8b67147ee250, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.234145] env[65788]: DEBUG oslo_vmware.api [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.505907] env[65788]: DEBUG nova.network.neutron [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Successfully created port: 82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 757.566024] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 757.570310] env[65788]: INFO nova.compute.manager [-] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Took 2.27 seconds to deallocate network for instance. [ 757.571813] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.572301] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.624502] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662194, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.626304] env[65788]: DEBUG nova.network.neutron [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 757.641847] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-391057a7-4e83-46e2-82ea-86c986ef1d0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.655983] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948c6445-8783-4200-896d-6fba801713fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.678178] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.678580] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.691033] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662195, 'name': Rename_Task, 'duration_secs': 0.16793} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.703341] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 757.704095] env[65788]: DEBUG nova.compute.manager [req-9c6ea331-7176-4ff4-bec8-9b594585feac req-d23637f7-1433-4849-8fd8-bb5ef15cce27 service nova] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Detach interface failed, port_id=709378e7-1b93-4d0c-9649-3fb90a38b091, reason: Instance 5b701040-025c-4246-ad54-f2cf478e998d could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 757.706896] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-905a03d9-cceb-4ac2-b65c-3d9e5ed83153 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.724449] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 757.724449] env[65788]: value = "task-4662198" [ 757.724449] env[65788]: _type = "Task" [ 757.724449] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.735335] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52de3e56-757f-ab76-6544-8b67147ee250, 'name': SearchDatastore_Task, 'duration_secs': 0.009996} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.735789] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.735962] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 757.737045] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a75e504-2efa-4e8a-8630-30b5898b49db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.744831] env[65788]: DEBUG oslo_vmware.api [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Task: {'id': task-4662197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147356} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.745083] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662198, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.748379] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 757.748624] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 757.749105] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.749105] env[65788]: INFO nova.compute.manager [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Took 1.51 seconds to destroy the instance on the hypervisor. [ 757.749400] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 757.749984] env[65788]: DEBUG nova.compute.manager [-] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 757.749984] env[65788]: DEBUG nova.network.neutron [-] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 757.749984] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.750534] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.750796] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.761246] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 757.761246] env[65788]: value = "task-4662199" [ 757.761246] env[65788]: _type = "Task" [ 757.761246] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.769130] env[65788]: WARNING neutronclient.v2_0.client [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.770140] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.770542] env[65788]: WARNING openstack [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.785826] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.832598] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 757.896975] env[65788]: DEBUG nova.network.neutron [-] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.083836] env[65788]: INFO nova.virt.block_device [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Booting with volume 0ec3463c-cdd6-4521-b484-dd20a13916e1 at /dev/sda [ 758.086866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.124155] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42c0c774-3647-45f5-af12-a92ac77fee5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.132712] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662194, 'name': CreateSnapshot_Task, 'duration_secs': 0.543266} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.133763] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 758.134697] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651e18de-83a8-4098-bb83-0262a872a436 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.144483] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57c347b-bd27-43b4-a426-92631a8e9231 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.191689] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e80cadc-ff8e-4ff0-bb41-c00f466f17ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.204115] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c940f9d3-2042-4731-96ab-b1280499d1fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.256727] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2aab75e-2a5f-48c4-8acc-f5b3fc8ed319 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.260133] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662198, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.270719] env[65788]: DEBUG nova.network.neutron [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Updating instance_info_cache with network_info: [{"id": "f116c9d8-4c98-4af9-8747-0c585d99d738", "address": "fa:16:3e:e3:a3:1e", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf116c9d8-4c", "ovs_interfaceid": "f116c9d8-4c98-4af9-8747-0c585d99d738", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.273770] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c05c4e4-417d-4bdb-96e8-96b8629f6be0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.284220] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662199, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.302273] env[65788]: DEBUG nova.virt.block_device [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updating existing volume attachment record: fdbde780-8864-4c8d-aa18-e7063123aeb2 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 758.364632] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6963359a-822c-4313-984d-aac2191dd968 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.374833] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e7de98c-d212-47dc-9a6f-9a936b05abb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.408294] env[65788]: INFO nova.compute.manager [-] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Took 1.29 seconds to deallocate network for instance. [ 758.410908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db252fab-445b-46a3-b8fb-afc7eb003278 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.421858] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f12b7fd-dc5a-4400-bcaa-156bf1a2bf1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.438498] env[65788]: DEBUG nova.compute.provider_tree [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.672933] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 758.673405] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dfb5a6be-ced8-4883-8e36-df7783c88f50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.683766] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 758.683766] env[65788]: value = "task-4662200" [ 758.683766] env[65788]: _type = "Task" [ 758.683766] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.695377] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662200, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.723037] env[65788]: DEBUG nova.network.neutron [-] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.740451] env[65788]: DEBUG oslo_vmware.api [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662198, 'name': PowerOnVM_Task, 'duration_secs': 0.891574} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.740451] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 758.740451] env[65788]: INFO nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Took 8.49 seconds to spawn the instance on the hypervisor. [ 758.740451] env[65788]: DEBUG nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 758.740611] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd3baca-3ca4-42f8-9881-69de9fecf395 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.773975] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.775032] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Instance network_info: |[{"id": "f116c9d8-4c98-4af9-8747-0c585d99d738", "address": "fa:16:3e:e3:a3:1e", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf116c9d8-4c", "ovs_interfaceid": "f116c9d8-4c98-4af9-8747-0c585d99d738", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 758.775032] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56326} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.775214] env[65788]: DEBUG oslo_concurrency.lockutils [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Acquired lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.775214] env[65788]: DEBUG nova.network.neutron [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Refreshing network info cache for port f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 758.776424] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:a3:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f116c9d8-4c98-4af9-8747-0c585d99d738', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.784464] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating folder: Project (c4cffe8dfc424459a90e316ac4ac905f). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.784464] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 758.784719] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.786046] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2375d96d-96f9-4b68-baf5-44d1d97d9bd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.787510] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb2dd0ef-f9d4-4d95-925f-1ee5e382ea76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.797269] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 758.797269] env[65788]: value = "task-4662202" [ 758.797269] env[65788]: _type = "Task" [ 758.797269] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.803413] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created folder: Project (c4cffe8dfc424459a90e316ac4ac905f) in parent group-v910111. [ 758.803645] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating folder: Instances. Parent ref: group-v910249. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.803939] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35778d42-2589-46bf-a8a8-de4df662d167 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.810673] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.817100] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created folder: Instances in parent group-v910249. [ 758.817295] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 758.817559] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.817898] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a9ec122-30d6-4b28-98bc-3acb391ee286 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.841190] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.841190] env[65788]: value = "task-4662204" [ 758.841190] env[65788]: _type = "Task" [ 758.841190] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.854441] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662204, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.919149] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.942572] env[65788]: DEBUG nova.scheduler.client.report [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 758.953760] env[65788]: DEBUG nova.compute.manager [req-cd92d81c-3077-4672-9ee5-9203ed33cb28 req-e4d4b2f6-e506-46d0-a019-f979e47e2695 service nova] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Received event network-vif-deleted-97cfb515-1e9e-4a23-bef7-86b855e9ab3f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 758.953952] env[65788]: DEBUG nova.compute.manager [req-cd92d81c-3077-4672-9ee5-9203ed33cb28 req-e4d4b2f6-e506-46d0-a019-f979e47e2695 service nova] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Received event network-vif-deleted-35fcd817-c75c-40df-beb8-b13fe638dd50 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 759.041302] env[65788]: DEBUG nova.compute.manager [req-6ad9c878-2bd6-413c-9291-ebea5c5f8cf7 req-8ae6e0c5-b92c-4d3f-8bee-985adda70d8a service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Received event network-vif-plugged-82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 759.042079] env[65788]: DEBUG oslo_concurrency.lockutils [req-6ad9c878-2bd6-413c-9291-ebea5c5f8cf7 req-8ae6e0c5-b92c-4d3f-8bee-985adda70d8a service nova] Acquiring lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.042079] env[65788]: DEBUG oslo_concurrency.lockutils [req-6ad9c878-2bd6-413c-9291-ebea5c5f8cf7 req-8ae6e0c5-b92c-4d3f-8bee-985adda70d8a service nova] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.042393] env[65788]: DEBUG oslo_concurrency.lockutils [req-6ad9c878-2bd6-413c-9291-ebea5c5f8cf7 req-8ae6e0c5-b92c-4d3f-8bee-985adda70d8a service nova] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.042472] env[65788]: DEBUG nova.compute.manager [req-6ad9c878-2bd6-413c-9291-ebea5c5f8cf7 req-8ae6e0c5-b92c-4d3f-8bee-985adda70d8a service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] No waiting events found dispatching network-vif-plugged-82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 759.042671] env[65788]: WARNING nova.compute.manager [req-6ad9c878-2bd6-413c-9291-ebea5c5f8cf7 req-8ae6e0c5-b92c-4d3f-8bee-985adda70d8a service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Received unexpected event network-vif-plugged-82d30609-6b15-4172-aef8-f9afa1ed72e9 for instance with vm_state building and task_state block_device_mapping. [ 759.143818] env[65788]: DEBUG nova.network.neutron [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Successfully updated port: 82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 759.195761] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662200, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.225606] env[65788]: INFO nova.compute.manager [-] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Took 1.48 seconds to deallocate network for instance. [ 759.262191] env[65788]: INFO nova.compute.manager [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Took 30.35 seconds to build instance. [ 759.277545] env[65788]: DEBUG oslo_concurrency.lockutils [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Acquiring lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.277744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Acquired lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.277959] env[65788]: DEBUG nova.network.neutron [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 759.285763] env[65788]: WARNING neutronclient.v2_0.client [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.286633] env[65788]: WARNING openstack [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.286984] env[65788]: WARNING openstack [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.310233] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084808} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.310645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 759.312086] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07b6a5a-97c5-449b-bb68-dbfad49cc455 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.333824] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 759.336730] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4242fd96-7d8f-4a41-8b27-2d859e4a00a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.362473] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662204, 'name': CreateVM_Task, 'duration_secs': 0.465623} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.364208] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 759.364576] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 759.364576] env[65788]: value = "task-4662205" [ 759.364576] env[65788]: _type = "Task" [ 759.364576] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.365529] env[65788]: WARNING neutronclient.v2_0.client [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.365529] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.365680] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.365966] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 759.366296] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-619c04d4-8b85-4249-88c8-f94a895460be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.378257] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662205, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.378760] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 759.378760] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522ce847-26e6-060b-d02c-f4e4402d80d4" [ 759.378760] env[65788]: _type = "Task" [ 759.378760] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.388442] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522ce847-26e6-060b-d02c-f4e4402d80d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.418133] env[65788]: WARNING openstack [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.418761] env[65788]: WARNING openstack [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.455630] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.189s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.456246] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 759.460578] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.431s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.460835] env[65788]: INFO nova.compute.claims [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.494902] env[65788]: WARNING neutronclient.v2_0.client [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.495596] env[65788]: WARNING openstack [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.495980] env[65788]: WARNING openstack [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.605767] env[65788]: DEBUG nova.network.neutron [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Updated VIF entry in instance network info cache for port f116c9d8-4c98-4af9-8747-0c585d99d738. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 759.605767] env[65788]: DEBUG nova.network.neutron [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Updating instance_info_cache with network_info: [{"id": "f116c9d8-4c98-4af9-8747-0c585d99d738", "address": "fa:16:3e:e3:a3:1e", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf116c9d8-4c", "ovs_interfaceid": "f116c9d8-4c98-4af9-8747-0c585d99d738", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 759.648327] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.648496] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquired lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.649043] env[65788]: DEBUG nova.network.neutron [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 759.695533] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662200, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.733364] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.766014] env[65788]: DEBUG oslo_concurrency.lockutils [None req-624d3005-35ce-4ab8-a0e6-1b7a838c58e2 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.781275] env[65788]: WARNING neutronclient.v2_0.client [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.782108] env[65788]: WARNING openstack [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.782503] env[65788]: WARNING openstack [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.882799] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662205, 'name': ReconfigVM_Task, 'duration_secs': 0.279279} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.886782] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Reconfigured VM instance instance-0000002a to attach disk [datastore2] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e/cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.887525] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e5d2403-aca2-4824-97b3-f1d989b2bf52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.890669] env[65788]: WARNING openstack [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.891512] env[65788]: WARNING openstack [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.908462] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522ce847-26e6-060b-d02c-f4e4402d80d4, 'name': SearchDatastore_Task, 'duration_secs': 0.017834} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.913418] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.913737] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 759.913980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.914166] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.914365] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 759.914758] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 759.914758] env[65788]: value = "task-4662206" [ 759.914758] env[65788]: _type = "Task" [ 759.914758] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.915263] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05c83914-ca6e-4356-9e4b-623a6116da12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.932513] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662206, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.934458] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 759.934710] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 759.935847] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0499885-4939-4f5c-ab44-640cdfd5e078 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.947958] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 759.947958] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c55a0-695e-ae1a-aeb7-0f7e081f7070" [ 759.947958] env[65788]: _type = "Task" [ 759.947958] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.962400] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c55a0-695e-ae1a-aeb7-0f7e081f7070, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.965950] env[65788]: DEBUG nova.compute.utils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 759.970315] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 759.970590] env[65788]: DEBUG nova.network.neutron [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 759.970878] env[65788]: WARNING neutronclient.v2_0.client [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.971229] env[65788]: WARNING neutronclient.v2_0.client [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.971832] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.972331] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 759.997798] env[65788]: WARNING neutronclient.v2_0.client [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 759.998636] env[65788]: WARNING openstack [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 759.999079] env[65788]: WARNING openstack [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.038046] env[65788]: DEBUG nova.policy [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9310d3b6bf5430289df41fc742584c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4cffe8dfc424459a90e316ac4ac905f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 760.114116] env[65788]: DEBUG oslo_concurrency.lockutils [req-dbbd21d3-9721-41d8-b141-d969d19f83b3 req-a7be685e-293a-4355-8114-186abab38f56 service nova] Releasing lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.131455] env[65788]: DEBUG nova.network.neutron [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Updating instance_info_cache with network_info: [{"id": "ea77fa15-d004-4f6d-bea0-02027383b034", "address": "fa:16:3e:eb:4d:86", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea77fa15-d0", "ovs_interfaceid": "ea77fa15-d004-4f6d-bea0-02027383b034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 760.153073] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.153594] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.196026] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662200, 'name': CloneVM_Task, 'duration_secs': 1.482995} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.196672] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Created linked-clone VM from snapshot [ 760.197500] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31450b0b-9385-43a9-8c17-44d78f15d900 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.204608] env[65788]: DEBUG nova.network.neutron [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 760.210725] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Uploading image c126b770-767b-4998-af13-f8bccb1538ed {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 760.234371] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.234777] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.253357] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 760.253357] env[65788]: value = "vm-910248" [ 760.253357] env[65788]: _type = "VirtualMachine" [ 760.253357] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 760.253839] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5c5c889d-3cd0-425b-a865-8a9fcf8135f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.262468] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease: (returnval){ [ 760.262468] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522132dc-c3a4-ff30-f73b-63018fa7e682" [ 760.262468] env[65788]: _type = "HttpNfcLease" [ 760.262468] env[65788]: } obtained for exporting VM: (result){ [ 760.262468] env[65788]: value = "vm-910248" [ 760.262468] env[65788]: _type = "VirtualMachine" [ 760.262468] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 760.262872] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the lease: (returnval){ [ 760.262872] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522132dc-c3a4-ff30-f73b-63018fa7e682" [ 760.262872] env[65788]: _type = "HttpNfcLease" [ 760.262872] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 760.267440] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 760.272757] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 760.272757] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522132dc-c3a4-ff30-f73b-63018fa7e682" [ 760.272757] env[65788]: _type = "HttpNfcLease" [ 760.272757] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 760.314462] env[65788]: WARNING neutronclient.v2_0.client [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 760.316028] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 760.316197] env[65788]: WARNING openstack [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 760.400357] env[65788]: DEBUG nova.network.neutron [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Successfully created port: 2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 760.417837] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 760.418426] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 760.418638] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 760.418786] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 760.418994] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 760.419123] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 760.419255] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 760.419458] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.419613] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 760.419876] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 760.419941] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 760.420127] env[65788]: DEBUG nova.virt.hardware [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 760.422752] env[65788]: DEBUG nova.network.neutron [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updating instance_info_cache with network_info: [{"id": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "address": "fa:16:3e:41:c1:f2", "network": {"id": "ee60d08c-870c-4eae-85be-e222dc1db4ba", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-128188116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81015cf240554f4b899ba703842162bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82d30609-6b", "ovs_interfaceid": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 760.427942] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df76a8e-c604-4e5e-af01-a5e5a6580fec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.446241] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ede3d4d-ac21-442e-bc4f-b0e99dc044ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.449804] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662206, 'name': Rename_Task, 'duration_secs': 0.211492} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.450763] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 760.454466] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60d0fb1a-96c0-4583-91b6-7419453dee9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.470616] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c55a0-695e-ae1a-aeb7-0f7e081f7070, 'name': SearchDatastore_Task, 'duration_secs': 0.016675} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.474172] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 760.476833] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 760.476833] env[65788]: value = "task-4662208" [ 760.476833] env[65788]: _type = "Task" [ 760.476833] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.478206] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-512397ff-db47-40cc-8fbb-b460d1da2eda {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.488782] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 760.488782] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52105384-1b36-23f2-edcd-30d7bc83503e" [ 760.488782] env[65788]: _type = "Task" [ 760.488782] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.492731] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662208, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.502706] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52105384-1b36-23f2-edcd-30d7bc83503e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.636275] env[65788]: DEBUG oslo_concurrency.lockutils [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Releasing lock "refresh_cache-f3c53f67-85bf-4c18-9313-75eb90862f78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.636556] env[65788]: DEBUG nova.compute.manager [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Inject network info {{(pid=65788) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7777}} [ 760.636884] env[65788]: DEBUG nova.compute.manager [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] network_info to inject: |[{"id": "ea77fa15-d004-4f6d-bea0-02027383b034", "address": "fa:16:3e:eb:4d:86", "network": {"id": "35750304-6b31-495b-8dde-d4211a8c713c", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1629426012-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06b3dfec7c0144928b4a59dbd27569b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2321dbbe-f64a-4253-a462-21676f8a278e", "external-id": "nsx-vlan-transportzone-714", "segmentation_id": 714, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapea77fa15-d0", "ovs_interfaceid": "ea77fa15-d004-4f6d-bea0-02027383b034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7778}} [ 760.643417] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Reconfiguring VM instance to set the machine id {{(pid=65788) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 760.647311] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-893ce373-0ddb-4148-b67a-f0576f0f9afe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.672569] env[65788]: DEBUG oslo_vmware.api [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Waiting for the task: (returnval){ [ 760.672569] env[65788]: value = "task-4662209" [ 760.672569] env[65788]: _type = "Task" [ 760.672569] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.688033] env[65788]: DEBUG oslo_vmware.api [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Task: {'id': task-4662209, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.780421] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 760.780421] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522132dc-c3a4-ff30-f73b-63018fa7e682" [ 760.780421] env[65788]: _type = "HttpNfcLease" [ 760.780421] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 760.783298] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 760.783298] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522132dc-c3a4-ff30-f73b-63018fa7e682" [ 760.783298] env[65788]: _type = "HttpNfcLease" [ 760.783298] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 760.784651] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ac3fbf-459c-4670-8e8c-169cf6126bec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.796130] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ee1ab5-5830-1c56-5ce5-ed20c42221df/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 760.796347] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ee1ab5-5830-1c56-5ce5-ed20c42221df/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 760.801909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 760.900737] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5a66ab8a-c3a9-4917-97a4-238e5153a69d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.932821] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Releasing lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 760.933350] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance network_info: |[{"id": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "address": "fa:16:3e:41:c1:f2", "network": {"id": "ee60d08c-870c-4eae-85be-e222dc1db4ba", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-128188116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81015cf240554f4b899ba703842162bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82d30609-6b", "ovs_interfaceid": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 760.933843] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:c1:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aef08290-001a-4ae8-aff0-1889e2211389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '82d30609-6b15-4172-aef8-f9afa1ed72e9', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 760.941675] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Creating folder: Project (81015cf240554f4b899ba703842162bc). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.952542] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd399085-748f-4309-950a-d59af8712df7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.974189] env[65788]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 760.974386] env[65788]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=65788) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 760.974812] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Folder already exists: Project (81015cf240554f4b899ba703842162bc). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 760.975040] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Creating folder: Instances. Parent ref: group-v910198. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.975565] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efd5af93-ce5c-4f81-b0fd-eb97a1c7aae2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.993452] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662208, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.997891] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Created folder: Instances in parent group-v910198. [ 760.998214] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 760.998585] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 760.999198] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bc1ae508-9c1e-4736-a738-3e1e0430315e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.023782] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52105384-1b36-23f2-edcd-30d7bc83503e, 'name': SearchDatastore_Task, 'duration_secs': 0.021875} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.024564] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.024823] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/9438ab56-1b4c-4778-a608-de319ab0ee43.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.025106] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-799b1b39-33e6-4716-ad47-1cc37559ea87 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.030970] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.030970] env[65788]: value = "task-4662212" [ 761.030970] env[65788]: _type = "Task" [ 761.030970] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.035748] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 761.035748] env[65788]: value = "task-4662213" [ 761.035748] env[65788]: _type = "Task" [ 761.035748] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.042732] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662212, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.051835] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662213, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.072492] env[65788]: DEBUG nova.compute.manager [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Received event network-changed-82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 761.072750] env[65788]: DEBUG nova.compute.manager [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Refreshing instance network info cache due to event network-changed-82d30609-6b15-4172-aef8-f9afa1ed72e9. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 761.073011] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Acquiring lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.073302] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Acquired lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 761.073519] env[65788]: DEBUG nova.network.neutron [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Refreshing network info cache for port 82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 761.186988] env[65788]: DEBUG oslo_vmware.api [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] Task: {'id': task-4662209, 'name': ReconfigVM_Task, 'duration_secs': 0.182941} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.187439] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-07e477cc-bd92-47d8-abdf-d9157ee40803 tempest-ServersAdminTestJSON-628619986 tempest-ServersAdminTestJSON-628619986-project-admin] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Reconfigured VM instance to set the machine id {{(pid=65788) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 761.241773] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e37401-aaa5-41c0-a516-c8224b2b67e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.251891] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6549fdef-ce35-4ee7-b3d5-f5361252a074 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.294340] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2986ed-dbc6-4c8f-a728-eb06fe2e0f79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.312516] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ff0189-e700-43fa-974c-0f836bac2c8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.336885] env[65788]: DEBUG nova.compute.provider_tree [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.487911] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 761.499062] env[65788]: DEBUG oslo_vmware.api [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662208, 'name': PowerOnVM_Task, 'duration_secs': 0.624584} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.499440] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 761.499776] env[65788]: DEBUG nova.compute.manager [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 761.500799] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6d60e6-29ab-44a3-80b3-4ecfd75546fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.525262] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 761.527513] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.527513] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 761.527513] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.527513] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 761.527513] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 761.527968] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.527968] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 761.527968] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 761.527968] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 761.527968] env[65788]: DEBUG nova.virt.hardware [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 761.528123] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef96c19-b107-415e-831a-7ede0f320054 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.543227] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f671fd8a-7ae1-4bad-8308-576bcabf753e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.552096] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662212, 'name': CreateVM_Task, 'duration_secs': 0.41895} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.553315] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.554557] env[65788]: WARNING neutronclient.v2_0.client [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.554557] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'fdbde780-8864-4c8d-aa18-e7063123aeb2', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910212', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'name': 'volume-0ec3463c-cdd6-4521-b484-dd20a13916e1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '263cd2fb-7eeb-4395-b1e9-b829940ebd35', 'attached_at': '', 'detached_at': '', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'serial': '0ec3463c-cdd6-4521-b484-dd20a13916e1'}, 'disk_bus': None, 'device_type': None, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65788) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 761.554741] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Root volume attach. Driver type: vmdk {{(pid=65788) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 761.555713] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cf6346-cd87-4bf4-95b0-47fdfd8d58c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.571195] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662213, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.576393] env[65788]: WARNING neutronclient.v2_0.client [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.577387] env[65788]: WARNING openstack [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.577912] env[65788]: WARNING openstack [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.591779] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8479ae-9bb0-46e9-8869-377847b5adf3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.600330] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc45f6b0-f5ac-47c9-bbba-eb90c76699e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.610242] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-0a0545af-9b53-4e3f-abe6-de793d887ccb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.620704] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 761.620704] env[65788]: value = "task-4662214" [ 761.620704] env[65788]: _type = "Task" [ 761.620704] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.634379] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662214, 'name': RelocateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.750823] env[65788]: WARNING openstack [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.753984] env[65788]: WARNING openstack [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.843059] env[65788]: DEBUG nova.scheduler.client.report [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.926368] env[65788]: WARNING neutronclient.v2_0.client [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 761.927289] env[65788]: WARNING openstack [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.927657] env[65788]: WARNING openstack [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.024459] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.051712] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662213, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612086} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.052302] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/9438ab56-1b4c-4778-a608-de319ab0ee43.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 762.052814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.054025] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86701133-d05a-4bbd-84b7-50b001492aaa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.063514] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 762.063514] env[65788]: value = "task-4662215" [ 762.063514] env[65788]: _type = "Task" [ 762.063514] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.073037] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662215, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.076157] env[65788]: DEBUG nova.network.neutron [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updated VIF entry in instance network info cache for port 82d30609-6b15-4172-aef8-f9afa1ed72e9. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 762.077197] env[65788]: DEBUG nova.network.neutron [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updating instance_info_cache with network_info: [{"id": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "address": "fa:16:3e:41:c1:f2", "network": {"id": "ee60d08c-870c-4eae-85be-e222dc1db4ba", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-128188116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81015cf240554f4b899ba703842162bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82d30609-6b", "ovs_interfaceid": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 762.097751] env[65788]: DEBUG nova.network.neutron [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Successfully updated port: 2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 762.134238] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662214, 'name': RelocateVM_Task, 'duration_secs': 0.031285} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.134491] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 762.134886] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910212', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'name': 'volume-0ec3463c-cdd6-4521-b484-dd20a13916e1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '263cd2fb-7eeb-4395-b1e9-b829940ebd35', 'attached_at': '', 'detached_at': '', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'serial': '0ec3463c-cdd6-4521-b484-dd20a13916e1'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 762.135790] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60cbd2d-e650-4df6-bb43-28f0bc491e62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.159651] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff29e696-436d-4404-905f-11313c7ddbba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.185759] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] volume-0ec3463c-cdd6-4521-b484-dd20a13916e1/volume-0ec3463c-cdd6-4521-b484-dd20a13916e1.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.186148] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1632d872-7ddd-4859-b448-b82936ebefcd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.210503] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 762.210503] env[65788]: value = "task-4662216" [ 762.210503] env[65788]: _type = "Task" [ 762.210503] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.223770] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662216, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.289181] env[65788]: DEBUG nova.compute.manager [req-47b4d2ef-92c2-4d5e-899d-ea34479118e6 req-cd0b17c5-32b0-4215-861e-f2d56e774432 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Received event network-vif-plugged-2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 762.289181] env[65788]: DEBUG oslo_concurrency.lockutils [req-47b4d2ef-92c2-4d5e-899d-ea34479118e6 req-cd0b17c5-32b0-4215-861e-f2d56e774432 service nova] Acquiring lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.289181] env[65788]: DEBUG oslo_concurrency.lockutils [req-47b4d2ef-92c2-4d5e-899d-ea34479118e6 req-cd0b17c5-32b0-4215-861e-f2d56e774432 service nova] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.289181] env[65788]: DEBUG oslo_concurrency.lockutils [req-47b4d2ef-92c2-4d5e-899d-ea34479118e6 req-cd0b17c5-32b0-4215-861e-f2d56e774432 service nova] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.289588] env[65788]: DEBUG nova.compute.manager [req-47b4d2ef-92c2-4d5e-899d-ea34479118e6 req-cd0b17c5-32b0-4215-861e-f2d56e774432 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] No waiting events found dispatching network-vif-plugged-2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 762.289900] env[65788]: WARNING nova.compute.manager [req-47b4d2ef-92c2-4d5e-899d-ea34479118e6 req-cd0b17c5-32b0-4215-861e-f2d56e774432 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Received unexpected event network-vif-plugged-2af7b9f3-396f-41c3-954c-e947acc02fd1 for instance with vm_state building and task_state spawning. [ 762.349196] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.890s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.349826] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 762.352898] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.886s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.353720] env[65788]: DEBUG nova.objects.instance [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'resources' on Instance uuid 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.577286] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662215, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122146} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.577854] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.578892] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddbbd59-14cc-4eaa-bb27-1d30b5f741b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.584532] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c335bac-84dc-48c5-895d-3c8c0d3ab464 req-cae2b040-451e-4b1b-b8c9-b7ac8f7e5c6f service nova] Releasing lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.603557] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/9438ab56-1b4c-4778-a608-de319ab0ee43.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.604299] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.604462] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.604931] env[65788]: DEBUG nova.network.neutron [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 762.605934] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7774df40-1fbb-49f4-b5c3-e8d7b374d21a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.621694] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.622094] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.636467] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 762.636467] env[65788]: value = "task-4662217" [ 762.636467] env[65788]: _type = "Task" [ 762.636467] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.652281] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662217, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.668720] env[65788]: DEBUG nova.network.neutron [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 762.688519] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.689226] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.723217] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662216, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.765045] env[65788]: WARNING neutronclient.v2_0.client [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.765769] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.766284] env[65788]: WARNING openstack [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.857522] env[65788]: DEBUG nova.compute.utils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 762.866345] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 762.866345] env[65788]: DEBUG nova.network.neutron [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 762.869293] env[65788]: WARNING neutronclient.v2_0.client [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.869293] env[65788]: WARNING neutronclient.v2_0.client [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 762.869293] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 762.870491] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 762.886290] env[65788]: DEBUG nova.network.neutron [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Updating instance_info_cache with network_info: [{"id": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "address": "fa:16:3e:1f:b4:80", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2af7b9f3-39", "ovs_interfaceid": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 762.956492] env[65788]: DEBUG nova.policy [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b0c1ce04a704a3eaf0a0d4dd09f7e08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9674e2a5c86b48db8c865a50331ab846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 763.146825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.146825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.146825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.146825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.147067] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.151037] env[65788]: INFO nova.compute.manager [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Terminating instance [ 763.162208] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662217, 'name': ReconfigVM_Task, 'duration_secs': 0.501374} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.165353] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/9438ab56-1b4c-4778-a608-de319ab0ee43.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.165353] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-278637de-9dee-42dc-9822-f83a757c35ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.173873] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 763.173873] env[65788]: value = "task-4662218" [ 763.173873] env[65788]: _type = "Task" [ 763.173873] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.186738] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662218, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.223644] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662216, 'name': ReconfigVM_Task, 'duration_secs': 0.654901} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.224732] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Reconfigured VM instance instance-0000002d to attach disk [datastore1] volume-0ec3463c-cdd6-4521-b484-dd20a13916e1/volume-0ec3463c-cdd6-4521-b484-dd20a13916e1.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.229190] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8cf0376-dbfa-422a-8b21-df9477b3235d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.248794] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 763.248794] env[65788]: value = "task-4662219" [ 763.248794] env[65788]: _type = "Task" [ 763.248794] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.259226] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662219, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.365848] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 763.376350] env[65788]: DEBUG nova.network.neutron [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Successfully created port: 9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 763.390030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.390145] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Instance network_info: |[{"id": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "address": "fa:16:3e:1f:b4:80", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2af7b9f3-39", "ovs_interfaceid": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 763.395201] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:b4:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2af7b9f3-396f-41c3-954c-e947acc02fd1', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.402275] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 763.402500] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 763.403225] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0349441f-f9df-4792-b1b1-d27434aeddbe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.432483] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.432483] env[65788]: value = "task-4662220" [ 763.432483] env[65788]: _type = "Task" [ 763.432483] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.444093] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662220, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.598250] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd62d88b-f804-49eb-9f44-8ca480aa772e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.608274] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde1f2bd-b5e4-4942-92d7-97b203d25a1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.646221] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b61a23-dd77-4b78-b237-6c319fb22fb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.657127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d040fcd4-4a0c-47e1-919f-3d5ecfa2d32d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.663415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "refresh_cache-cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.663679] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquired lock "refresh_cache-cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.663869] env[65788]: DEBUG nova.network.neutron [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 763.676619] env[65788]: DEBUG nova.compute.provider_tree [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.690655] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662218, 'name': Rename_Task, 'duration_secs': 0.220967} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.691094] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 763.691927] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93b424e3-d3da-4c05-8fcb-51b44c2265d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.701264] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 763.701264] env[65788]: value = "task-4662221" [ 763.701264] env[65788]: _type = "Task" [ 763.701264] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.712629] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.762249] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662219, 'name': ReconfigVM_Task, 'duration_secs': 0.194975} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.762584] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910212', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'name': 'volume-0ec3463c-cdd6-4521-b484-dd20a13916e1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '263cd2fb-7eeb-4395-b1e9-b829940ebd35', 'attached_at': '', 'detached_at': '', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'serial': '0ec3463c-cdd6-4521-b484-dd20a13916e1'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 763.763164] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a0ca3a6-c42c-494f-8300-38d720a2d306 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.772120] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 763.772120] env[65788]: value = "task-4662222" [ 763.772120] env[65788]: _type = "Task" [ 763.772120] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.776095] env[65788]: INFO nova.compute.manager [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Rebuilding instance [ 763.787478] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662222, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.822373] env[65788]: DEBUG nova.compute.manager [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 763.822555] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7782fbb-de8d-4687-bae2-518863f858bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.948882] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662220, 'name': CreateVM_Task, 'duration_secs': 0.404071} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.949439] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 763.950263] env[65788]: WARNING neutronclient.v2_0.client [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 763.950850] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.951152] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.951637] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 763.952059] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0214d404-4818-4003-a828-6b2dea2c253d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.959656] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 763.959656] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d2bb4d-cb0f-4d32-1f41-effe0fe99bea" [ 763.959656] env[65788]: _type = "Task" [ 763.959656] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.973069] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d2bb4d-cb0f-4d32-1f41-effe0fe99bea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.166951] env[65788]: WARNING neutronclient.v2_0.client [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.167704] env[65788]: WARNING openstack [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.168185] env[65788]: WARNING openstack [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.182482] env[65788]: DEBUG nova.scheduler.client.report [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.206381] env[65788]: DEBUG nova.network.neutron [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 764.215830] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662221, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.281418] env[65788]: DEBUG nova.network.neutron [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 764.286190] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662222, 'name': Rename_Task, 'duration_secs': 0.185025} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.286693] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 764.286975] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c39d035a-25e4-4c34-8807-e893d80d5de1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.297341] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 764.297341] env[65788]: value = "task-4662223" [ 764.297341] env[65788]: _type = "Task" [ 764.297341] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.310405] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.321890] env[65788]: DEBUG nova.compute.manager [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Received event network-changed-2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 764.322228] env[65788]: DEBUG nova.compute.manager [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Refreshing instance network info cache due to event network-changed-2af7b9f3-396f-41c3-954c-e947acc02fd1. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 764.322386] env[65788]: DEBUG oslo_concurrency.lockutils [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Acquiring lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.322460] env[65788]: DEBUG oslo_concurrency.lockutils [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Acquired lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 764.322607] env[65788]: DEBUG nova.network.neutron [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Refreshing network info cache for port 2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 764.385315] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 764.414920] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 764.415254] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 764.415472] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 764.416162] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 764.416162] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 764.416162] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 764.416345] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.416469] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 764.416639] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 764.416799] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 764.416967] env[65788]: DEBUG nova.virt.hardware [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 764.418018] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e441e79d-c18f-48bc-93b3-646743ed8e73 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.430924] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc38e03-1ba4-4fea-85d8-0bcb07f9899e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.473227] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d2bb4d-cb0f-4d32-1f41-effe0fe99bea, 'name': SearchDatastore_Task, 'duration_secs': 0.014392} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.473227] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.473227] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 764.473227] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.473494] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 764.473494] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 764.473494] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87584816-479a-4027-ba0f-2ac283c834d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.484270] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 764.484472] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 764.485271] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bac866bf-d9e5-425a-b573-e72463dd0612 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.492322] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 764.492322] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ffdd71-0162-b089-15a3-2a4cd09d01a9" [ 764.492322] env[65788]: _type = "Task" [ 764.492322] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.502191] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ffdd71-0162-b089-15a3-2a4cd09d01a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.688529] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.336s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.691148] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.339s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 764.691378] env[65788]: DEBUG nova.objects.instance [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 764.715797] env[65788]: DEBUG oslo_vmware.api [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662221, 'name': PowerOnVM_Task, 'duration_secs': 0.634567} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.716678] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 764.716678] env[65788]: INFO nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Took 9.31 seconds to spawn the instance on the hypervisor. [ 764.716855] env[65788]: DEBUG nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 764.717679] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31e0fed-2ce7-4480-bec9-e2effd4aa518 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.741837] env[65788]: INFO nova.scheduler.client.report [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted allocations for instance 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c [ 764.787367] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Releasing lock "refresh_cache-cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.787956] env[65788]: DEBUG nova.compute.manager [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 764.788292] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.789781] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73d296f-7d9f-42e8-9d84-f0ee6ad72e2d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.800428] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 764.805822] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01f9b12a-6597-4baf-afdd-e6498b3efdef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.818381] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.820631] env[65788]: DEBUG oslo_vmware.api [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 764.820631] env[65788]: value = "task-4662224" [ 764.820631] env[65788]: _type = "Task" [ 764.820631] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.826677] env[65788]: WARNING neutronclient.v2_0.client [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 764.826677] env[65788]: WARNING openstack [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.826677] env[65788]: WARNING openstack [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.842616] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 764.842923] env[65788]: DEBUG oslo_vmware.api [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662224, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.843179] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31290586-53bc-41b4-8252-a4a20986c725 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.852020] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 764.852020] env[65788]: value = "task-4662225" [ 764.852020] env[65788]: _type = "Task" [ 764.852020] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.864219] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.973508] env[65788]: WARNING openstack [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.973935] env[65788]: WARNING openstack [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.004086] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ffdd71-0162-b089-15a3-2a4cd09d01a9, 'name': SearchDatastore_Task, 'duration_secs': 0.01731} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.004954] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-debdba70-fec1-40f2-ae39-60868794dc41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.013163] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 765.013163] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f54a9b-8b80-6b9f-c895-78644c3a1104" [ 765.013163] env[65788]: _type = "Task" [ 765.013163] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.025318] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f54a9b-8b80-6b9f-c895-78644c3a1104, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.054839] env[65788]: DEBUG nova.network.neutron [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Successfully updated port: 9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 765.118820] env[65788]: WARNING neutronclient.v2_0.client [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.119585] env[65788]: WARNING openstack [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.119952] env[65788]: WARNING openstack [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.222231] env[65788]: DEBUG nova.network.neutron [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Updated VIF entry in instance network info cache for port 2af7b9f3-396f-41c3-954c-e947acc02fd1. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 765.222651] env[65788]: DEBUG nova.network.neutron [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Updating instance_info_cache with network_info: [{"id": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "address": "fa:16:3e:1f:b4:80", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2af7b9f3-39", "ovs_interfaceid": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 765.237954] env[65788]: INFO nova.compute.manager [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Took 33.60 seconds to build instance. [ 765.250904] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7dd4af81-579e-48dd-b47b-648220e4132e tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "4b7f1c2f-2b1e-4d24-814b-c8095d875e3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.548s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.310810] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.331751] env[65788]: DEBUG oslo_vmware.api [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662224, 'name': PowerOffVM_Task, 'duration_secs': 0.219869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.332067] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.332288] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 765.332568] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-088045f0-a691-45e8-ac5e-1c87d44e72f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.360590] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 765.360973] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 765.361320] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Deleting the datastore file [datastore2] cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.364649] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a156849c-39d0-45ac-ab60-9c9182f7aeb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.366672] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662225, 'name': PowerOffVM_Task, 'duration_secs': 0.300457} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.366978] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.367277] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.368594] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f083fe3-3ac7-450c-ae8c-cd1720fe99de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.373351] env[65788]: DEBUG oslo_vmware.api [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for the task: (returnval){ [ 765.373351] env[65788]: value = "task-4662227" [ 765.373351] env[65788]: _type = "Task" [ 765.373351] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.379328] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 765.380670] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8657f1ba-82ea-46d9-95c7-d5ede02aaa62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.388122] env[65788]: DEBUG oslo_vmware.api [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.469289] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 765.469289] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 765.469602] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleting the datastore file [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.469818] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c23de2a-8455-40d1-9888-c9a3de70b1fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.479820] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 765.479820] env[65788]: value = "task-4662229" [ 765.479820] env[65788]: _type = "Task" [ 765.479820] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.491982] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662229, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.525589] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f54a9b-8b80-6b9f-c895-78644c3a1104, 'name': SearchDatastore_Task, 'duration_secs': 0.013079} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.526293] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.527012] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d78d36e6-a542-4ba8-9e29-b8a7b28c559e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 765.527690] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a83ef804-b8ba-43f9-8be2-e121437666f6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.537317] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 765.537317] env[65788]: value = "task-4662230" [ 765.537317] env[65788]: _type = "Task" [ 765.537317] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.548221] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.557358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.557575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 765.557775] env[65788]: DEBUG nova.network.neutron [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 765.701467] env[65788]: DEBUG oslo_concurrency.lockutils [None req-276cad28-89f5-4168-a5de-0b0e11400428 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.702554] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.069s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.702811] env[65788]: DEBUG nova.objects.instance [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lazy-loading 'resources' on Instance uuid 3f20bd75-98ab-4798-834a-c0ffc3c6146d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.727088] env[65788]: DEBUG oslo_concurrency.lockutils [req-17d36b20-3a1f-44fd-b64f-46955fa8ea4d req-c60ad866-0150-4f58-bdf2-c70d6c022b27 service nova] Releasing lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 765.739861] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c6a19a5d-12ef-4721-b9b4-734fcd971af3 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.147s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.812584] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.890591] env[65788]: DEBUG oslo_vmware.api [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Task: {'id': task-4662227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1832} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.891375] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 765.891591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 765.891809] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.891989] env[65788]: INFO nova.compute.manager [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 765.892362] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 765.892603] env[65788]: DEBUG nova.compute.manager [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 765.892718] env[65788]: DEBUG nova.network.neutron [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 765.892985] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 765.895136] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.896966] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.924374] env[65788]: DEBUG nova.network.neutron [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 765.924374] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.002548] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662229, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274506} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.003121] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 766.003458] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 766.003843] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.064029] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.064029] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.070913] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662230, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.126516] env[65788]: DEBUG nova.network.neutron [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 766.149790] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.150613] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.234408] env[65788]: WARNING neutronclient.v2_0.client [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 766.235310] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 766.235764] env[65788]: WARNING openstack [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 766.244921] env[65788]: DEBUG nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 766.317270] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.364379] env[65788]: DEBUG nova.network.neutron [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance_info_cache with network_info: [{"id": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "address": "fa:16:3e:7c:87:d4", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e5e0e29-c7", "ovs_interfaceid": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.429296] env[65788]: DEBUG nova.network.neutron [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.555752] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662230, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677353} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.555988] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d78d36e6-a542-4ba8-9e29-b8a7b28c559e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 766.557098] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.557098] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5eebf729-0621-400b-a552-3deb699e827e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.566204] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 766.566204] env[65788]: value = "task-4662231" [ 766.566204] env[65788]: _type = "Task" [ 766.566204] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.583752] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.776091] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.780926] env[65788]: DEBUG nova.compute.manager [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Received event network-vif-plugged-9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 766.781216] env[65788]: DEBUG oslo_concurrency.lockutils [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Acquiring lock "f75000e6-f3bf-4951-a541-3aa185339054-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.781417] env[65788]: DEBUG oslo_concurrency.lockutils [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Lock "f75000e6-f3bf-4951-a541-3aa185339054-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.785221] env[65788]: DEBUG oslo_concurrency.lockutils [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Lock "f75000e6-f3bf-4951-a541-3aa185339054-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.785221] env[65788]: DEBUG nova.compute.manager [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] No waiting events found dispatching network-vif-plugged-9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 766.785221] env[65788]: WARNING nova.compute.manager [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Received unexpected event network-vif-plugged-9e5e0e29-c770-41e3-9766-2be8e6f3dc96 for instance with vm_state building and task_state spawning. [ 766.785221] env[65788]: DEBUG nova.compute.manager [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Received event network-changed-9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 766.785221] env[65788]: DEBUG nova.compute.manager [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Refreshing instance network info cache due to event network-changed-9e5e0e29-c770-41e3-9766-2be8e6f3dc96. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 766.785475] env[65788]: DEBUG oslo_concurrency.lockutils [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Acquiring lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.819790] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.869202] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.869202] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Instance network_info: |[{"id": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "address": "fa:16:3e:7c:87:d4", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e5e0e29-c7", "ovs_interfaceid": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 766.869492] env[65788]: DEBUG oslo_concurrency.lockutils [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Acquired lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.869956] env[65788]: DEBUG nova.network.neutron [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Refreshing network info cache for port 9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 766.871510] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:87:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9e5e0e29-c770-41e3-9766-2be8e6f3dc96', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 766.879930] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 766.881624] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 766.881624] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1df54f2-bab4-42da-9cfb-023105cf438c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.899928] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b5b445-96a3-44ad-8292-db2b02c627cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.912602] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f87d32-be0f-4b5d-bef8-9973158c9674 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.916051] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 766.916051] env[65788]: value = "task-4662232" [ 766.916051] env[65788]: _type = "Task" [ 766.916051] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.950993] env[65788]: INFO nova.compute.manager [-] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Took 1.06 seconds to deallocate network for instance. [ 766.954463] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0779579-5fc4-4c3f-85a3-e8dd21b03457 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.964282] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662232, 'name': CreateVM_Task} progress is 15%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.973581] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6024879-76df-46f0-b006-b1e8c0cb187d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.988850] env[65788]: DEBUG nova.compute.provider_tree [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.055602] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 767.055918] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 767.056359] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 767.056608] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 767.056777] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 767.056934] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 767.057164] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.058622] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 767.058622] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 767.058622] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 767.058622] env[65788]: DEBUG nova.virt.hardware [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 767.058900] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e65e182-a7ce-4585-8036-7857bd34635e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.072663] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55896ed8-c4b8-435a-82d5-aed7ac8a2adc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.083850] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08236} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.092332] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 767.092986] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:9f:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b0ec4f3-be92-493e-a8ed-6af0f0bb844c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.101546] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 767.101934] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d588f57-8eaa-4fc1-b71b-f6e870a9d7f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.105152] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.107092] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1db7631f-7930-4cec-a795-33233ea38dc4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.150086] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d78d36e6-a542-4ba8-9e29-b8a7b28c559e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 767.152789] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6471b148-359e-4e74-beb5-0e2fead98f00 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.172945] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.172945] env[65788]: value = "task-4662233" [ 767.172945] env[65788]: _type = "Task" [ 767.172945] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.183031] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 767.183031] env[65788]: value = "task-4662234" [ 767.183031] env[65788]: _type = "Task" [ 767.183031] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.187715] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662233, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.199789] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662234, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.317958] env[65788]: DEBUG oslo_vmware.api [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662223, 'name': PowerOnVM_Task, 'duration_secs': 2.602171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.317958] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.318360] env[65788]: INFO nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Took 6.90 seconds to spawn the instance on the hypervisor. [ 767.318360] env[65788]: DEBUG nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 767.319181] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d73521a-11f7-4ed4-bc76-c36770d546a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.382186] env[65788]: WARNING neutronclient.v2_0.client [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.382589] env[65788]: WARNING openstack [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.382794] env[65788]: WARNING openstack [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.427828] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662232, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.465971] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.491983] env[65788]: DEBUG nova.scheduler.client.report [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 767.501399] env[65788]: WARNING openstack [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.501831] env[65788]: WARNING openstack [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.574742] env[65788]: WARNING neutronclient.v2_0.client [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 767.575830] env[65788]: WARNING openstack [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 767.575830] env[65788]: WARNING openstack [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 767.683603] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662233, 'name': CreateVM_Task, 'duration_secs': 0.508478} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.687808] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 767.690254] env[65788]: DEBUG nova.network.neutron [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updated VIF entry in instance network info cache for port 9e5e0e29-c770-41e3-9766-2be8e6f3dc96. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 767.690657] env[65788]: DEBUG nova.network.neutron [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance_info_cache with network_info: [{"id": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "address": "fa:16:3e:7c:87:d4", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e5e0e29-c7", "ovs_interfaceid": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 767.692258] env[65788]: WARNING neutronclient.v2_0.client [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 767.692742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.692814] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.694313] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 767.694313] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-932327a4-60f4-4bf1-b95a-2c7362dcef90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.699768] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662234, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.704420] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 767.704420] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52573c61-2f3d-95e6-eb61-2fce30297afc" [ 767.704420] env[65788]: _type = "Task" [ 767.704420] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.714094] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52573c61-2f3d-95e6-eb61-2fce30297afc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.719188] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquiring lock "54ad3472-0cca-4dca-91f2-8c343f706926" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.719526] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "54ad3472-0cca-4dca-91f2-8c343f706926" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.719776] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquiring lock "54ad3472-0cca-4dca-91f2-8c343f706926-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.720015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "54ad3472-0cca-4dca-91f2-8c343f706926-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.720485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "54ad3472-0cca-4dca-91f2-8c343f706926-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.724970] env[65788]: INFO nova.compute.manager [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Terminating instance [ 767.840378] env[65788]: INFO nova.compute.manager [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Took 36.07 seconds to build instance. [ 767.933294] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662232, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.997547] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.295s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.000952] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.531s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.002885] env[65788]: INFO nova.compute.claims [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.030496] env[65788]: INFO nova.scheduler.client.report [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Deleted allocations for instance 3f20bd75-98ab-4798-834a-c0ffc3c6146d [ 768.197378] env[65788]: DEBUG oslo_concurrency.lockutils [req-f55f96d8-8925-417a-9014-739b10a0a51c req-265d4b3b-815e-4339-91dd-8f06c561467f service nova] Releasing lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.197823] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662234, 'name': ReconfigVM_Task, 'duration_secs': 0.590848} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.198707] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Reconfigured VM instance instance-0000002e to attach disk [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d78d36e6-a542-4ba8-9e29-b8a7b28c559e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 768.198908] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5344609-c9a5-47f3-a9dd-b1402fd98737 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.207398] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 768.207398] env[65788]: value = "task-4662235" [ 768.207398] env[65788]: _type = "Task" [ 768.207398] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.217938] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52573c61-2f3d-95e6-eb61-2fce30297afc, 'name': SearchDatastore_Task, 'duration_secs': 0.013111} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.218941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.219342] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.219543] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.219696] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.220201] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.220201] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c7407dd-8d68-4a09-9e95-f51b60bef410 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.226080] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662235, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.232634] env[65788]: DEBUG nova.compute.manager [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 768.232890] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.233222] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c0e3bd93-8b2b-4312-9ebc-b926e1d01aec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.236728] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.236919] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.238137] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50ebb3d3-191c-44e8-8628-19c7a2715949 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.246655] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 768.246655] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f4e6f1-c720-72f4-9b53-808d2700f4e1" [ 768.246655] env[65788]: _type = "Task" [ 768.246655] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.252234] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 768.252234] env[65788]: value = "task-4662236" [ 768.252234] env[65788]: _type = "Task" [ 768.252234] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.261492] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f4e6f1-c720-72f4-9b53-808d2700f4e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.265375] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.300779] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ee1ab5-5830-1c56-5ce5-ed20c42221df/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 768.301474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569fa941-79a3-4706-b2fa-ca301112fce4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.309118] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ee1ab5-5830-1c56-5ce5-ed20c42221df/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 768.309372] env[65788]: ERROR oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ee1ab5-5830-1c56-5ce5-ed20c42221df/disk-0.vmdk due to incomplete transfer. [ 768.309652] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-db212194-1621-478b-b600-91f5c4b4d0f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.321591] env[65788]: DEBUG oslo_vmware.rw_handles [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ee1ab5-5830-1c56-5ce5-ed20c42221df/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 768.321591] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Uploaded image c126b770-767b-4998-af13-f8bccb1538ed to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 768.322914] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 768.323295] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-491cce6b-f56f-4503-97c6-3a041b611978 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.332775] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 768.332775] env[65788]: value = "task-4662237" [ 768.332775] env[65788]: _type = "Task" [ 768.332775] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.344324] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d63faf2-579b-4ed0-b28d-213d0b47bd3f tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.520s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.344699] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662237, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.431024] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662232, 'name': CreateVM_Task, 'duration_secs': 1.447604} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.431987] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 768.434020] env[65788]: WARNING neutronclient.v2_0.client [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.434020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.434020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.434020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 768.434020] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d48556fb-aafc-4e36-b4d2-7c7ac765493d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.441029] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 768.441029] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff3877-c3d4-4606-57e0-0a36e3b7ce9f" [ 768.441029] env[65788]: _type = "Task" [ 768.441029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.451823] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff3877-c3d4-4606-57e0-0a36e3b7ce9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.539966] env[65788]: DEBUG oslo_concurrency.lockutils [None req-96d855db-1ebb-419c-8681-f993979802bf tempest-ServersAaction247Test-1870518301 tempest-ServersAaction247Test-1870518301-project-member] Lock "3f20bd75-98ab-4798-834a-c0ffc3c6146d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.570s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.605776] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.606145] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.723026] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662235, 'name': Rename_Task, 'duration_secs': 0.304388} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.723026] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 768.723475] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c83164d7-606a-4aca-95b7-fd650d415f5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.732215] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 768.732215] env[65788]: value = "task-4662238" [ 768.732215] env[65788]: _type = "Task" [ 768.732215] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.744112] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662238, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.763564] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f4e6f1-c720-72f4-9b53-808d2700f4e1, 'name': SearchDatastore_Task, 'duration_secs': 0.036895} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.764763] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0035006e-d2b2-4d04-a053-7b154b2f0a7e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.770277] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662236, 'name': PowerOffVM_Task, 'duration_secs': 0.298013} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.771112] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.771228] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 768.771422] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910189', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'name': 'volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '54ad3472-0cca-4dca-91f2-8c343f706926', 'attached_at': '', 'detached_at': '', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'serial': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 768.772247] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e13c56a-c81d-4bd6-92cf-d85eb169adfd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.777201] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 768.777201] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520dfca7-2a66-bd62-1bec-3211e7407f08" [ 768.777201] env[65788]: _type = "Task" [ 768.777201] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.799702] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee5d65f-3de6-41af-aca4-28d82a5da2be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.807429] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520dfca7-2a66-bd62-1bec-3211e7407f08, 'name': SearchDatastore_Task, 'duration_secs': 0.022025} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.808494] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.808870] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.809287] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c177311-d27e-481b-a20b-faa56440149f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.815570] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aceccbee-4b97-4b80-985f-49782c04bef6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.821935] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 768.821935] env[65788]: value = "task-4662239" [ 768.821935] env[65788]: _type = "Task" [ 768.821935] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.845397] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4625c43e-bc3c-455e-a726-ad6f1dee6344 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.848252] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 768.860578] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.874977] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] The volume has not been displaced from its original location: [datastore2] volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea/volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 768.880339] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Reconfiguring VM instance instance-00000024 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 768.880744] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662237, 'name': Destroy_Task, 'duration_secs': 0.526993} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.881297] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3aa5436-619b-4b09-ad65-1ae5d55143f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.894567] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Destroyed the VM [ 768.894842] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 768.895164] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-085fd5de-6df7-4797-8b57-e361ddb4fa8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.904261] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 768.904261] env[65788]: value = "task-4662240" [ 768.904261] env[65788]: _type = "Task" [ 768.904261] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.906696] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 768.906696] env[65788]: value = "task-4662241" [ 768.906696] env[65788]: _type = "Task" [ 768.906696] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.930871] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662240, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.931191] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662241, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.954149] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff3877-c3d4-4606-57e0-0a36e3b7ce9f, 'name': SearchDatastore_Task, 'duration_secs': 0.013328} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.955297] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 768.955575] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.955918] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.956089] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 768.956278] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.956881] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a4fd416-9643-4611-bf60-00bbb6a92714 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.968513] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.968795] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 768.969679] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27e52632-aa63-49f5-ade2-99a08487c3e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.976561] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 768.976561] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c47e5d-1117-fd4d-e39d-7bf417c89445" [ 768.976561] env[65788]: _type = "Task" [ 768.976561] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.988714] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c47e5d-1117-fd4d-e39d-7bf417c89445, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.249218] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662238, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.360562] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662239, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.382769] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.431724] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662240, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.445626] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662241, 'name': ReconfigVM_Task, 'duration_secs': 0.195758} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.445626] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Reconfigured VM instance instance-00000024 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 769.453237] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbde0f95-f0a1-49b6-aacb-d699d1fcf5d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.473814] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 769.473814] env[65788]: value = "task-4662242" [ 769.473814] env[65788]: _type = "Task" [ 769.473814] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.489488] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662242, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.494226] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c47e5d-1117-fd4d-e39d-7bf417c89445, 'name': SearchDatastore_Task, 'duration_secs': 0.010996} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.495182] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2f98539-5d56-4393-9246-23c27eabe1ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.502660] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 769.502660] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525de6be-4eea-3cdf-216a-2a405f2baa09" [ 769.502660] env[65788]: _type = "Task" [ 769.502660] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.519911] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525de6be-4eea-3cdf-216a-2a405f2baa09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.540131] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 769.541087] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 769.718866] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7c1482-0196-47fa-b95c-55113ce00e39 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.728119] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180b2169-4a28-45f2-97aa-dd45467ec57c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.766138] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f0d944-44cd-4b58-a4ed-4f7198779d1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.770237] env[65788]: DEBUG nova.compute.manager [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Received event network-changed-82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 769.770549] env[65788]: DEBUG nova.compute.manager [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Refreshing instance network info cache due to event network-changed-82d30609-6b15-4172-aef8-f9afa1ed72e9. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 769.770613] env[65788]: DEBUG oslo_concurrency.lockutils [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Acquiring lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.770752] env[65788]: DEBUG oslo_concurrency.lockutils [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Acquired lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.770906] env[65788]: DEBUG nova.network.neutron [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Refreshing network info cache for port 82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 769.778703] env[65788]: DEBUG oslo_vmware.api [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662238, 'name': PowerOnVM_Task, 'duration_secs': 0.700563} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.778703] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 769.778703] env[65788]: INFO nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Took 8.29 seconds to spawn the instance on the hypervisor. [ 769.778703] env[65788]: DEBUG nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 769.778703] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e6ba5b-8a42-4ccb-a037-d5ccaac7c6b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.785133] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d301b5-9777-42ae-bc09-27a5c423dab1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.805901] env[65788]: DEBUG nova.compute.provider_tree [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.851222] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662239, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670065} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.851526] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.851765] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.852064] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6dc79a6-a8ad-46ea-a83d-25b302e9b0b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.861707] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 769.861707] env[65788]: value = "task-4662243" [ 769.861707] env[65788]: _type = "Task" [ 769.861707] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.876763] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662243, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.920778] env[65788]: DEBUG oslo_vmware.api [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662240, 'name': RemoveSnapshot_Task, 'duration_secs': 0.655531} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.921077] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 769.921369] env[65788]: INFO nova.compute.manager [None req-bef65cbd-75ff-4d56-99a0-9af456d3d464 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Took 13.82 seconds to snapshot the instance on the hypervisor. [ 769.985605] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662242, 'name': ReconfigVM_Task, 'duration_secs': 0.183077} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.985927] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910189', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'name': 'volume-54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '54ad3472-0cca-4dca-91f2-8c343f706926', 'attached_at': '', 'detached_at': '', 'volume_id': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea', 'serial': '54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 769.986297] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.987039] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2d7024-c58d-442e-8faf-ebfeed8a10f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.999244] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 769.999244] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8ff5c79-3a2e-48ba-9d91-86fc20c68f13 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.024312] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525de6be-4eea-3cdf-216a-2a405f2baa09, 'name': SearchDatastore_Task, 'duration_secs': 0.01686} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.026067] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.026067] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] f75000e6-f3bf-4951-a541-3aa185339054/f75000e6-f3bf-4951-a541-3aa185339054.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 770.029179] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b771e266-80de-4f07-92bc-843e1a34bc5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.051800] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 770.051800] env[65788]: value = "task-4662245" [ 770.051800] env[65788]: _type = "Task" [ 770.051800] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.057337] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.058348] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.058348] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.058710] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.058710] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.058825] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.059240] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 770.059240] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 770.065998] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662245, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.082043] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 770.082043] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 770.082043] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Deleting the datastore file [datastore2] 54ad3472-0cca-4dca-91f2-8c343f706926 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.082378] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97561c5e-a2a0-4691-af3e-fecae7c00442 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.091604] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for the task: (returnval){ [ 770.091604] env[65788]: value = "task-4662246" [ 770.091604] env[65788]: _type = "Task" [ 770.091604] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.103443] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.277425] env[65788]: WARNING neutronclient.v2_0.client [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.278215] env[65788]: WARNING openstack [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.278594] env[65788]: WARNING openstack [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.302627] env[65788]: INFO nova.compute.manager [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Took 37.43 seconds to build instance. [ 770.308982] env[65788]: DEBUG nova.scheduler.client.report [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 770.376609] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662243, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073665} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.376984] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.377869] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57ac99b-2fe9-493e-9d04-ccd4dc5b390e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.405802] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 770.411068] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f14659a-a693-47e3-b04f-7db5f74c6f71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.437195] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 770.437195] env[65788]: value = "task-4662247" [ 770.437195] env[65788]: _type = "Task" [ 770.437195] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.452942] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662247, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.465985] env[65788]: WARNING openstack [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.466469] env[65788]: WARNING openstack [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.542052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "99f2cc13-a559-4379-8332-d432e8bac36b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.542052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "99f2cc13-a559-4379-8332-d432e8bac36b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.542052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "99f2cc13-a559-4379-8332-d432e8bac36b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.542052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "99f2cc13-a559-4379-8332-d432e8bac36b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.542228] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "99f2cc13-a559-4379-8332-d432e8bac36b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.543815] env[65788]: INFO nova.compute.manager [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Terminating instance [ 770.554034] env[65788]: WARNING neutronclient.v2_0.client [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.554745] env[65788]: WARNING openstack [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.555200] env[65788]: WARNING openstack [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.571277] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 770.580549] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662245, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.604034] env[65788]: DEBUG oslo_vmware.api [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Task: {'id': task-4662246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185311} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.604283] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.604665] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 770.604665] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.604872] env[65788]: INFO nova.compute.manager [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Took 2.37 seconds to destroy the instance on the hypervisor. [ 770.605101] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 770.605323] env[65788]: DEBUG nova.compute.manager [-] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 770.605458] env[65788]: DEBUG nova.network.neutron [-] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 770.605694] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.606293] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.606550] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.660033] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 770.695293] env[65788]: DEBUG nova.network.neutron [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updated VIF entry in instance network info cache for port 82d30609-6b15-4172-aef8-f9afa1ed72e9. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 770.696626] env[65788]: DEBUG nova.network.neutron [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updating instance_info_cache with network_info: [{"id": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "address": "fa:16:3e:41:c1:f2", "network": {"id": "ee60d08c-870c-4eae-85be-e222dc1db4ba", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-128188116-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.223", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "81015cf240554f4b899ba703842162bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aef08290-001a-4ae8-aff0-1889e2211389", "external-id": "nsx-vlan-transportzone-389", "segmentation_id": 389, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap82d30609-6b", "ovs_interfaceid": "82d30609-6b15-4172-aef8-f9afa1ed72e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.804887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6fb996bb-b0c3-41fa-87d3-160a6e6b3722 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.035s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.814843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.814s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.815905] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 770.819144] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 27.248s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 770.948173] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662247, 'name': ReconfigVM_Task, 'duration_secs': 0.325289} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.948173] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Reconfigured VM instance instance-00000010 to attach disk [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.948796] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-274d74bf-4185-4a75-af92-00a56cbe2b3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.956371] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 770.956371] env[65788]: value = "task-4662248" [ 770.956371] env[65788]: _type = "Task" [ 770.956371] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.967635] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662248, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.054832] env[65788]: DEBUG nova.compute.manager [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 771.054832] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.054832] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442c9013-1036-42c2-9fb4-735e2938d84e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.063828] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 771.064235] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99e8199c-13f2-4191-b295-fcabbc845252 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.078425] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662245, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.647936} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.078721] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] f75000e6-f3bf-4951-a541-3aa185339054/f75000e6-f3bf-4951-a541-3aa185339054.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 771.078933] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.079270] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fecf0abf-806e-4c5c-918d-1baabcf385ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.086845] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 771.086845] env[65788]: value = "task-4662250" [ 771.086845] env[65788]: _type = "Task" [ 771.086845] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.096487] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.147526] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 771.148023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 771.148023] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleting the datastore file [datastore2] 99f2cc13-a559-4379-8332-d432e8bac36b {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.148285] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f908d0c-416b-4a32-b446-8f7b173ff8dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.156014] env[65788]: DEBUG oslo_vmware.api [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 771.156014] env[65788]: value = "task-4662251" [ 771.156014] env[65788]: _type = "Task" [ 771.156014] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.166019] env[65788]: DEBUG oslo_vmware.api [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.200339] env[65788]: DEBUG oslo_concurrency.lockutils [req-2ff5204b-d72e-4086-8b9e-73b7ce8570c5 req-5da4bac8-3bd8-43b1-9f6f-efcb55ce4173 service nova] Releasing lock "refresh_cache-263cd2fb-7eeb-4395-b1e9-b829940ebd35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.283204] env[65788]: DEBUG nova.compute.manager [req-5f9538f3-1e01-44e7-bac2-f3e39090ac2d req-4e65d06c-e66b-48dd-9a75-9eefaeb3383f service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Received event network-vif-deleted-2117f600-f365-489d-9475-bf0afdc8f512 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 771.283428] env[65788]: INFO nova.compute.manager [req-5f9538f3-1e01-44e7-bac2-f3e39090ac2d req-4e65d06c-e66b-48dd-9a75-9eefaeb3383f service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Neutron deleted interface 2117f600-f365-489d-9475-bf0afdc8f512; detaching it from the instance and deleting it from the info cache [ 771.283582] env[65788]: DEBUG nova.network.neutron [req-5f9538f3-1e01-44e7-bac2-f3e39090ac2d req-4e65d06c-e66b-48dd-9a75-9eefaeb3383f service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.304066] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "37471e59-1809-4df3-8c40-20921d04d18e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.304341] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "37471e59-1809-4df3-8c40-20921d04d18e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.307585] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 771.326456] env[65788]: DEBUG nova.compute.utils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 771.334207] env[65788]: INFO nova.compute.claims [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.340433] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 771.340669] env[65788]: DEBUG nova.network.neutron [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 771.341104] env[65788]: WARNING neutronclient.v2_0.client [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.341424] env[65788]: WARNING neutronclient.v2_0.client [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.342038] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.343238] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.436425] env[65788]: DEBUG nova.policy [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afdd193345c74bcb9e28467267c1fbd5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df050bf0673c41f483ccec937a9d3ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 771.468286] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662248, 'name': Rename_Task, 'duration_secs': 0.157444} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.468286] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.468286] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb786f91-8b40-4a0f-acab-63776f92e845 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.476522] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 771.476522] env[65788]: value = "task-4662252" [ 771.476522] env[65788]: _type = "Task" [ 771.476522] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.485661] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.598441] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106426} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.598678] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.599723] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a73aea5-cb8d-4d64-a5a9-7c8f1d9bbcb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.624455] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] f75000e6-f3bf-4951-a541-3aa185339054/f75000e6-f3bf-4951-a541-3aa185339054.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.624623] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78bc367e-ec6c-4dd7-bfb9-2175e847f853 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.645974] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 771.645974] env[65788]: value = "task-4662253" [ 771.645974] env[65788]: _type = "Task" [ 771.645974] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.655923] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.666811] env[65788]: DEBUG oslo_vmware.api [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160383} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.667103] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.667450] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 771.667640] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.667805] env[65788]: INFO nova.compute.manager [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Took 0.62 seconds to destroy the instance on the hypervisor. [ 771.668084] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 771.668427] env[65788]: DEBUG nova.compute.manager [-] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 771.668484] env[65788]: DEBUG nova.network.neutron [-] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 771.668740] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.669387] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 771.669652] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 771.692536] env[65788]: DEBUG nova.network.neutron [-] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.748307] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 771.754773] env[65788]: DEBUG nova.network.neutron [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Successfully created port: 3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 771.788842] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6030e26c-26d4-48b0-8acd-c6223ada0c1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.800978] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1237a4fa-245c-407a-bf65-dee39cfb0d3d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.848152] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 771.850143] env[65788]: DEBUG nova.compute.manager [req-5f9538f3-1e01-44e7-bac2-f3e39090ac2d req-4e65d06c-e66b-48dd-9a75-9eefaeb3383f service nova] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Detach interface failed, port_id=2117f600-f365-489d-9475-bf0afdc8f512, reason: Instance 54ad3472-0cca-4dca-91f2-8c343f706926 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 771.851750] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.853377] env[65788]: INFO nova.compute.resource_tracker [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating resource usage from migration 7fa860c2-3230-490e-9d69-7e7474adafa0 [ 771.898644] env[65788]: INFO nova.compute.manager [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Rescuing [ 771.898977] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.899175] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.899396] env[65788]: DEBUG nova.network.neutron [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 771.988709] env[65788]: DEBUG oslo_vmware.api [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662252, 'name': PowerOnVM_Task, 'duration_secs': 0.502291} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.991108] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 771.991383] env[65788]: DEBUG nova.compute.manager [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 771.992822] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72965ab-0dbf-4a6f-b283-bc8af30c21bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.162790] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662253, 'name': ReconfigVM_Task, 'duration_secs': 0.307057} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.163107] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Reconfigured VM instance instance-0000002f to attach disk [datastore1] f75000e6-f3bf-4951-a541-3aa185339054/f75000e6-f3bf-4951-a541-3aa185339054.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.163821] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f8890963-784c-402e-b4f9-91c1ec8116b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.174508] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 772.174508] env[65788]: value = "task-4662254" [ 772.174508] env[65788]: _type = "Task" [ 772.174508] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.189210] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662254, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.195398] env[65788]: INFO nova.compute.manager [-] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Took 1.59 seconds to deallocate network for instance. [ 772.328869] env[65788]: DEBUG nova.compute.manager [req-b0f4e4ac-6c8a-4b6a-846a-bf8f982e8c35 req-20770147-9b87-44e8-ad2e-7ad3121fbec0 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Received event network-vif-deleted-a490dde6-7a99-4085-9de3-e81116f08231 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 772.329124] env[65788]: INFO nova.compute.manager [req-b0f4e4ac-6c8a-4b6a-846a-bf8f982e8c35 req-20770147-9b87-44e8-ad2e-7ad3121fbec0 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Neutron deleted interface a490dde6-7a99-4085-9de3-e81116f08231; detaching it from the instance and deleting it from the info cache [ 772.329351] env[65788]: DEBUG nova.network.neutron [req-b0f4e4ac-6c8a-4b6a-846a-bf8f982e8c35 req-20770147-9b87-44e8-ad2e-7ad3121fbec0 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.407189] env[65788]: WARNING neutronclient.v2_0.client [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.408028] env[65788]: WARNING openstack [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.408318] env[65788]: WARNING openstack [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.474707] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34987ff-e8bb-4661-bfd3-a576289c6268 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.483830] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade8038a-f2e0-49f1-b9f3-081a45d673a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.523961] env[65788]: DEBUG nova.network.neutron [-] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.531539] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1532c3cd-1ccb-48f0-9ada-0e0e1d68655c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.535137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.542026] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145a6f83-6c38-46ed-af86-539ab75a0c39 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.558485] env[65788]: DEBUG nova.compute.provider_tree [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.560594] env[65788]: WARNING openstack [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.560982] env[65788]: WARNING openstack [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.646227] env[65788]: WARNING neutronclient.v2_0.client [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 772.646897] env[65788]: WARNING openstack [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.647260] env[65788]: WARNING openstack [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.686282] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662254, 'name': Rename_Task, 'duration_secs': 0.156562} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.686373] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.686641] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7582aea-52e7-41eb-bb01-f46c2c3b1c34 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.694855] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 772.694855] env[65788]: value = "task-4662255" [ 772.694855] env[65788]: _type = "Task" [ 772.694855] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.711274] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662255, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.761748] env[65788]: INFO nova.compute.manager [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Took 0.57 seconds to detach 1 volumes for instance. [ 772.764473] env[65788]: DEBUG nova.compute.manager [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Deleting volume: 54d80a3e-5ce3-4d33-81c8-cec9bc49f0ea {{(pid=65788) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3320}} [ 772.777551] env[65788]: DEBUG nova.network.neutron [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Updating instance_info_cache with network_info: [{"id": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "address": "fa:16:3e:1f:b4:80", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2af7b9f3-39", "ovs_interfaceid": "2af7b9f3-396f-41c3-954c-e947acc02fd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.832366] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-871d534d-5d43-44bb-a297-bb2aa60d1798 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.843639] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc5e87a-c302-4503-8935-f9e50c031dfd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.856876] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 772.900134] env[65788]: DEBUG nova.compute.manager [req-b0f4e4ac-6c8a-4b6a-846a-bf8f982e8c35 req-20770147-9b87-44e8-ad2e-7ad3121fbec0 service nova] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Detach interface failed, port_id=a490dde6-7a99-4085-9de3-e81116f08231, reason: Instance 99f2cc13-a559-4379-8332-d432e8bac36b could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 772.910484] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 772.910830] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 772.911030] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 772.911361] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 772.911541] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 772.911720] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 772.912061] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.912294] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 772.912510] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 772.912696] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 772.912871] env[65788]: DEBUG nova.virt.hardware [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 772.914900] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111eee8b-ba4c-4a68-9e11-7ec98b522877 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.925772] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba9eb07-7e13-4b49-8d06-8bc97153efbf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.036090] env[65788]: INFO nova.compute.manager [-] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Took 1.37 seconds to deallocate network for instance. [ 773.068937] env[65788]: DEBUG nova.scheduler.client.report [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 773.206969] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662255, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.280254] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-d78d36e6-a542-4ba8-9e29-b8a7b28c559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.331539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.441585] env[65788]: DEBUG nova.network.neutron [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Successfully updated port: 3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 773.476023] env[65788]: DEBUG nova.compute.manager [req-11b0ec3b-84c3-4d05-92f0-d994c0199a16 req-96d44308-561c-4101-8545-2dc543230192 service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Received event network-vif-plugged-3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 773.476023] env[65788]: DEBUG oslo_concurrency.lockutils [req-11b0ec3b-84c3-4d05-92f0-d994c0199a16 req-96d44308-561c-4101-8545-2dc543230192 service nova] Acquiring lock "5c7a1693-62f2-454e-9406-0b4a132ebf25-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.476023] env[65788]: DEBUG oslo_concurrency.lockutils [req-11b0ec3b-84c3-4d05-92f0-d994c0199a16 req-96d44308-561c-4101-8545-2dc543230192 service nova] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.476023] env[65788]: DEBUG oslo_concurrency.lockutils [req-11b0ec3b-84c3-4d05-92f0-d994c0199a16 req-96d44308-561c-4101-8545-2dc543230192 service nova] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.476023] env[65788]: DEBUG nova.compute.manager [req-11b0ec3b-84c3-4d05-92f0-d994c0199a16 req-96d44308-561c-4101-8545-2dc543230192 service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] No waiting events found dispatching network-vif-plugged-3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 773.476447] env[65788]: WARNING nova.compute.manager [req-11b0ec3b-84c3-4d05-92f0-d994c0199a16 req-96d44308-561c-4101-8545-2dc543230192 service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Received unexpected event network-vif-plugged-3a66ab74-5f68-4ab5-939e-de3cc3e69db2 for instance with vm_state building and task_state spawning. [ 773.543387] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.575014] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.756s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.575240] env[65788]: INFO nova.compute.manager [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Migrating [ 773.582990] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.069s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.583257] env[65788]: DEBUG nova.objects.instance [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lazy-loading 'resources' on Instance uuid 0548399d-ec7a-4d0c-b436-9ef93c4e2d37 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 773.705638] env[65788]: DEBUG oslo_vmware.api [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662255, 'name': PowerOnVM_Task, 'duration_secs': 0.603355} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.705935] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.706155] env[65788]: INFO nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Took 9.32 seconds to spawn the instance on the hypervisor. [ 773.706336] env[65788]: DEBUG nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 773.707126] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef22f361-936e-4f78-b7e0-abf0f15ed766 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.945923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "refresh_cache-5c7a1693-62f2-454e-9406-0b4a132ebf25" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.945923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired lock "refresh_cache-5c7a1693-62f2-454e-9406-0b4a132ebf25" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.945923] env[65788]: DEBUG nova.network.neutron [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 774.095899] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.096116] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 774.096301] env[65788]: DEBUG nova.network.neutron [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 774.230196] env[65788]: INFO nova.compute.manager [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Rebuilding instance [ 774.232844] env[65788]: INFO nova.compute.manager [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Took 39.23 seconds to build instance. [ 774.285496] env[65788]: DEBUG nova.compute.manager [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 774.286406] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2aec28-4871-4164-bbe9-875718bce231 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.452059] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.452527] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.494075] env[65788]: DEBUG nova.network.neutron [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 774.514727] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.515152] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.591613] env[65788]: WARNING neutronclient.v2_0.client [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.592743] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.593098] env[65788]: WARNING openstack [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.603058] env[65788]: WARNING neutronclient.v2_0.client [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.604791] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.604791] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.642990] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbd2083-f238-4d39-a2cb-1c974ff92b2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.654167] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f3630c-fadb-42ed-b4da-52cb927eb978 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.697174] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa93f10-89eb-44a5-8d6f-defbb1057f6a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.709038] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3246616-e3cc-404e-bf07-78f347a318d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.712876] env[65788]: DEBUG nova.network.neutron [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Updating instance_info_cache with network_info: [{"id": "3a66ab74-5f68-4ab5-939e-de3cc3e69db2", "address": "fa:16:3e:73:dc:1a", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.198", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66ab74-5f", "ovs_interfaceid": "3a66ab74-5f68-4ab5-939e-de3cc3e69db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.729884] env[65788]: DEBUG nova.compute.provider_tree [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.738112] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e179f5e1-1240-45b3-9cc2-d17540510aaa tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.675s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.770616] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.771026] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.821984] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 774.823550] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd93334d-5749-4112-abbe-9615c686437f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.835022] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 774.835022] env[65788]: value = "task-4662257" [ 774.835022] env[65788]: _type = "Task" [ 774.835022] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.842750] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662257, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.844095] env[65788]: WARNING neutronclient.v2_0.client [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.845413] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 774.845413] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 774.950732] env[65788]: DEBUG nova.network.neutron [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 775.216263] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Releasing lock "refresh_cache-5c7a1693-62f2-454e-9406-0b4a132ebf25" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 775.216650] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Instance network_info: |[{"id": "3a66ab74-5f68-4ab5-939e-de3cc3e69db2", "address": "fa:16:3e:73:dc:1a", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.198", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66ab74-5f", "ovs_interfaceid": "3a66ab74-5f68-4ab5-939e-de3cc3e69db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 775.217148] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:dc:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a66ab74-5f68-4ab5-939e-de3cc3e69db2', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.224905] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 775.225751] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 775.225845] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7bf07e83-7dc7-471c-b6da-16507fa3a221 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.241031] env[65788]: DEBUG nova.scheduler.client.report [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 775.244523] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 775.253348] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.253348] env[65788]: value = "task-4662258" [ 775.253348] env[65788]: _type = "Task" [ 775.253348] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.263256] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662258, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.309664] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 775.309664] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28ef48df-3b47-47c5-9760-8849f88eb65e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.317545] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 775.317545] env[65788]: value = "task-4662259" [ 775.317545] env[65788]: _type = "Task" [ 775.317545] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.329023] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662259, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.346579] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662257, 'name': PowerOffVM_Task, 'duration_secs': 0.243683} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.346985] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.348275] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c72795-16c8-48a2-9ae7-2e0466e13a97 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.371239] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbff5c26-5e54-4815-bb9a-590a46980257 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.408940] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 775.409323] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dac90c6d-eaff-40f7-baa4-91067aa46281 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.420020] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 775.420020] env[65788]: value = "task-4662260" [ 775.420020] env[65788]: _type = "Task" [ 775.420020] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.431290] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 775.431290] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.431787] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.432072] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 775.432350] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.432686] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9deed58e-547f-4f72-9e53-f415638c894c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.444316] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.444833] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 775.445905] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3bd0510-a9b6-4435-92f7-aca5a59e28c6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.453813] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 775.455953] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 775.455953] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae2589-7a8a-12ce-8810-b737e8afce2e" [ 775.455953] env[65788]: _type = "Task" [ 775.455953] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.470032] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae2589-7a8a-12ce-8810-b737e8afce2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.659069] env[65788]: DEBUG nova.compute.manager [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Received event network-changed-3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 775.659343] env[65788]: DEBUG nova.compute.manager [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Refreshing instance network info cache due to event network-changed-3a66ab74-5f68-4ab5-939e-de3cc3e69db2. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 775.659459] env[65788]: DEBUG oslo_concurrency.lockutils [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Acquiring lock "refresh_cache-5c7a1693-62f2-454e-9406-0b4a132ebf25" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.659551] env[65788]: DEBUG oslo_concurrency.lockutils [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Acquired lock "refresh_cache-5c7a1693-62f2-454e-9406-0b4a132ebf25" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 775.659706] env[65788]: DEBUG nova.network.neutron [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Refreshing network info cache for port 3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 775.749093] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.166s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.754034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.427s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.755633] env[65788]: INFO nova.compute.claims [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.770292] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662258, 'name': CreateVM_Task, 'duration_secs': 0.410614} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.770466] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 775.771014] env[65788]: WARNING neutronclient.v2_0.client [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.771446] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.771607] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 775.772207] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 775.772295] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10b3d834-6b1a-4c74-845d-312eeeafa118 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.777182] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.778734] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 775.778734] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529395e1-fe45-1ae5-a55b-9b84de1ac5ff" [ 775.778734] env[65788]: _type = "Task" [ 775.778734] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.779557] env[65788]: INFO nova.scheduler.client.report [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Deleted allocations for instance 0548399d-ec7a-4d0c-b436-9ef93c4e2d37 [ 775.795621] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529395e1-fe45-1ae5-a55b-9b84de1ac5ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.828298] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662259, 'name': PowerOffVM_Task, 'duration_secs': 0.223008} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.828460] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.828684] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.829495] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66787c54-fe9f-4f51-a32c-907f597f0395 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.837625] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.837874] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbb7123d-48be-48c9-938a-f38b7c2fe780 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.909120] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 775.909407] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 775.909596] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleting the datastore file [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 775.909904] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fdc3cf8-78a1-4f04-b66c-8f6b5fd5269c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.918414] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 775.918414] env[65788]: value = "task-4662262" [ 775.918414] env[65788]: _type = "Task" [ 775.918414] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.930539] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662262, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.973399] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae2589-7a8a-12ce-8810-b737e8afce2e, 'name': SearchDatastore_Task, 'duration_secs': 0.018477} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.974240] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37c078b7-983f-44c3-bf1b-b06a64142ab7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.981820] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 775.981820] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520f6ec5-cde8-8502-6dfa-bd591f8afb11" [ 775.981820] env[65788]: _type = "Task" [ 775.981820] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.993264] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520f6ec5-cde8-8502-6dfa-bd591f8afb11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.165088] env[65788]: WARNING neutronclient.v2_0.client [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.165088] env[65788]: WARNING openstack [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.165088] env[65788]: WARNING openstack [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.296360] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529395e1-fe45-1ae5-a55b-9b84de1ac5ff, 'name': SearchDatastore_Task, 'duration_secs': 0.017706} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.297025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77814509-e080-40e8-a0ae-b0a316b51493 tempest-ServersListShow296Test-1312221363 tempest-ServersListShow296Test-1312221363-project-member] Lock "0548399d-ec7a-4d0c-b436-9ef93c4e2d37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.499s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.298915] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 776.298915] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.298915] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.298915] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 776.299335] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.300629] env[65788]: WARNING openstack [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.301011] env[65788]: WARNING openstack [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.308991] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-743600ee-7211-4a0c-b19e-a8c401b638bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.323547] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.324782] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 776.324840] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91c12e00-1a0e-407e-9874-1b3248e81420 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.334464] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 776.334464] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5257baec-4c00-dd63-6ebb-1d15b8fe00de" [ 776.334464] env[65788]: _type = "Task" [ 776.334464] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.349162] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5257baec-4c00-dd63-6ebb-1d15b8fe00de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.386625] env[65788]: WARNING neutronclient.v2_0.client [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 776.387175] env[65788]: WARNING openstack [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.387530] env[65788]: WARNING openstack [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.431488] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662262, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208121} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.432445] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.432842] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.433467] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.494165] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520f6ec5-cde8-8502-6dfa-bd591f8afb11, 'name': SearchDatastore_Task, 'duration_secs': 0.023805} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.494524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 776.494814] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. {{(pid=65788) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 776.495269] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7183e938-accb-49fb-a1e1-d26283ad660d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.500596] env[65788]: DEBUG nova.network.neutron [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Updated VIF entry in instance network info cache for port 3a66ab74-5f68-4ab5-939e-de3cc3e69db2. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 776.501321] env[65788]: DEBUG nova.network.neutron [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Updating instance_info_cache with network_info: [{"id": "3a66ab74-5f68-4ab5-939e-de3cc3e69db2", "address": "fa:16:3e:73:dc:1a", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.198", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a66ab74-5f", "ovs_interfaceid": "3a66ab74-5f68-4ab5-939e-de3cc3e69db2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 776.506028] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 776.506028] env[65788]: value = "task-4662263" [ 776.506028] env[65788]: _type = "Task" [ 776.506028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.514215] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662263, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.698128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "670f59c3-10b0-431f-a8b0-ef6f82a938be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.698128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.698128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "670f59c3-10b0-431f-a8b0-ef6f82a938be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 776.698128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 776.698685] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 776.699508] env[65788]: INFO nova.compute.manager [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Terminating instance [ 776.853813] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5257baec-4c00-dd63-6ebb-1d15b8fe00de, 'name': SearchDatastore_Task, 'duration_secs': 0.025101} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.857685] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dab55c1f-b483-40d3-99c8-d43c6058d574 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.864248] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 776.864248] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525d01b4-e9a4-9470-dc9b-34488075c4ea" [ 776.864248] env[65788]: _type = "Task" [ 776.864248] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.876054] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525d01b4-e9a4-9470-dc9b-34488075c4ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.974433] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf73f5b-cdad-4ac3-b804-40eb14473cfa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.006364] env[65788]: DEBUG oslo_concurrency.lockutils [req-22dce992-5863-4e31-9999-631fbdce412b req-5cd3b519-21c1-47fc-bfc9-0b15280061ce service nova] Releasing lock "refresh_cache-5c7a1693-62f2-454e-9406-0b4a132ebf25" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 777.007081] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 777.032746] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662263, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.203999] env[65788]: DEBUG nova.compute.manager [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 777.204455] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 777.205624] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9438bc-be9d-4572-95eb-645c687caaaf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.218584] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 777.222054] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-947cd417-f261-4d7f-8605-f81f839073f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.232565] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 777.232565] env[65788]: value = "task-4662264" [ 777.232565] env[65788]: _type = "Task" [ 777.232565] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.249625] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.384022] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525d01b4-e9a4-9470-dc9b-34488075c4ea, 'name': SearchDatastore_Task, 'duration_secs': 0.012432} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.384022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 777.384022] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5c7a1693-62f2-454e-9406-0b4a132ebf25/5c7a1693-62f2-454e-9406-0b4a132ebf25.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 777.384022] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-892e3eff-b2d5-4a33-9c8d-56b6d2ebd508 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.396087] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 777.396087] env[65788]: value = "task-4662265" [ 777.396087] env[65788]: _type = "Task" [ 777.396087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.405680] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662265, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.417992] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176db817-f46f-4114-8c7e-e7dc21e9ed59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.430289] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090faaa9-f6ea-49eb-a33d-d90d746f666d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.436089] env[65788]: DEBUG nova.compute.manager [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 777.471971] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffea9739-c0a5-4db1-bdee-9dc023189d70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.483481] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ecc35f-6ac1-4d68-8e05-2914372b9c66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.504096] env[65788]: DEBUG nova.compute.provider_tree [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.509087] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 777.509087] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 777.509087] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 777.509087] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 777.509321] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 777.509321] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 777.509321] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.509321] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 777.509321] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 777.509552] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 777.509552] env[65788]: DEBUG nova.virt.hardware [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 777.510495] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3cb262-ddff-4c29-94df-11ec7dda5370 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.514577] env[65788]: DEBUG nova.scheduler.client.report [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 777.522787] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 777.523472] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb79f86f-3630-4907-b72d-df2c25eaec92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.529505] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8779618e-2a82-43bd-a534-e45f93d20257 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.540156] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662263, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.540156] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 777.540156] env[65788]: value = "task-4662266" [ 777.540156] env[65788]: _type = "Task" [ 777.540156] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.540156] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. [ 777.540659] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b622f0-94e5-4bc0-9efa-ceaae5d4ea8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.554435] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:9f:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2321dbbe-f64a-4253-a462-21676f8a278e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b0ec4f3-be92-493e-a8ed-6af0f0bb844c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.561840] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 777.563072] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.563311] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-437e787f-aa42-4540-a46e-d60b1f83a99a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.604962] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.605709] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8084b253-3ab6-49b4-80a2-36ac8aa3aad9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.619009] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 777.619268] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 777.624324] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.624324] env[65788]: value = "task-4662267" [ 777.624324] env[65788]: _type = "Task" [ 777.624324] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.629751] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 777.629751] env[65788]: value = "task-4662268" [ 777.629751] env[65788]: _type = "Task" [ 777.629751] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.637277] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662267, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.644917] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662268, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.744297] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662264, 'name': PowerOffVM_Task, 'duration_secs': 0.251912} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.744297] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 777.744696] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 777.746775] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48b51f71-02ad-49fc-9544-c8ae5e3a5b1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.821918] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 777.822221] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 777.822467] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Deleting the datastore file [datastore2] 670f59c3-10b0-431f-a8b0-ef6f82a938be {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.822817] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79d2560b-6c83-4857-8c27-f633c624afd1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.834542] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for the task: (returnval){ [ 777.834542] env[65788]: value = "task-4662270" [ 777.834542] env[65788]: _type = "Task" [ 777.834542] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.844296] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662270, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.911161] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662265, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.965873] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.023139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.023871] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 778.028185] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.808s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 778.030408] env[65788]: INFO nova.compute.claims [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.127568] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 778.127876] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 778.128083] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 778.128279] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 778.128445] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 778.128725] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 778.128814] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.128989] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 778.129200] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 778.129369] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 778.129543] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 778.137993] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7885ceba-2c10-411f-87c0-77d24f2a1aad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.157825] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662267, 'name': CreateVM_Task, 'duration_secs': 0.452586} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.161639] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 778.161939] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662268, 'name': ReconfigVM_Task, 'duration_secs': 0.3595} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.162353] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 778.162353] env[65788]: value = "task-4662271" [ 778.162353] env[65788]: _type = "Task" [ 778.162353] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.162748] env[65788]: WARNING neutronclient.v2_0.client [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 778.163228] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.163496] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.163690] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 778.164014] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Reconfigured VM instance instance-0000002e to attach disk [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.164366] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-940ac6ba-4f99-46c9-ab59-e15f542a25fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.166672] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea79c1a9-15f3-4a97-a867-8ece9a4cfeb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.204271] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdb123d6-b6cc-45b8-8ed1-ff88f389063c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.215399] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662271, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.215885] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 778.215885] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c70870-00c8-5b92-1670-ff41be9d99fa" [ 778.215885] env[65788]: _type = "Task" [ 778.215885] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.223583] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 778.223583] env[65788]: value = "task-4662272" [ 778.223583] env[65788]: _type = "Task" [ 778.223583] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.232035] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c70870-00c8-5b92-1670-ff41be9d99fa, 'name': SearchDatastore_Task, 'duration_secs': 0.020437} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.233070] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.233324] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.233560] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.233700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.233872] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.234171] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d32a3b4-8453-4059-9ad5-b5235ccfa5d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.243361] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662272, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.247020] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.247020] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.247422] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-679119a7-e8d0-445c-bc17-972a1e71d62e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.254114] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 778.254114] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52594767-60e3-8095-c12c-c87f8007ef10" [ 778.254114] env[65788]: _type = "Task" [ 778.254114] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.263511] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52594767-60e3-8095-c12c-c87f8007ef10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.346371] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662270, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.410493] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662265, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.808825} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.410756] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5c7a1693-62f2-454e-9406-0b4a132ebf25/5c7a1693-62f2-454e-9406-0b4a132ebf25.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 778.410968] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.411272] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8821e18-00d9-4590-87e6-68113cd211f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.419406] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 778.419406] env[65788]: value = "task-4662273" [ 778.419406] env[65788]: _type = "Task" [ 778.419406] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.429754] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662273, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.538253] env[65788]: DEBUG nova.compute.utils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 778.541809] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 778.542024] env[65788]: DEBUG nova.network.neutron [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 778.542368] env[65788]: WARNING neutronclient.v2_0.client [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.542698] env[65788]: WARNING neutronclient.v2_0.client [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.544059] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.544321] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.609533] env[65788]: DEBUG nova.policy [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd68c01d2bad24cc1b6305141afa0e5ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41f69a1056bc454e83bd9561cf12c506', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 778.675298] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662271, 'name': ReconfigVM_Task, 'duration_secs': 0.22875} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.675609] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 778.740143] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662272, 'name': ReconfigVM_Task, 'duration_secs': 0.438481} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.740443] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 778.740795] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c5b747c-37cd-4de7-bf89-ab975cdbc10a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.748576] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 778.748576] env[65788]: value = "task-4662274" [ 778.748576] env[65788]: _type = "Task" [ 778.748576] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.758542] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.767461] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52594767-60e3-8095-c12c-c87f8007ef10, 'name': SearchDatastore_Task, 'duration_secs': 0.028561} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.768368] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8895ec4-841a-49bb-bce5-8060f59a41d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.774179] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 778.774179] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243eb71-f25d-8fcb-c9bf-4a3386c7b214" [ 778.774179] env[65788]: _type = "Task" [ 778.774179] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.783428] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243eb71-f25d-8fcb-c9bf-4a3386c7b214, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.845086] env[65788]: DEBUG oslo_vmware.api [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Task: {'id': task-4662270, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.794145} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.845274] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 778.845430] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 778.845607] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.845814] env[65788]: INFO nova.compute.manager [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Took 1.64 seconds to destroy the instance on the hypervisor. [ 778.846062] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 778.846265] env[65788]: DEBUG nova.compute.manager [-] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 778.846357] env[65788]: DEBUG nova.network.neutron [-] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 778.846611] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.847326] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.847433] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.929948] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 778.934923] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662273, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06779} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.935770] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.937021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be3a8dc-44c8-4709-9608-96b2cc9b4cb4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.963947] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 5c7a1693-62f2-454e-9406-0b4a132ebf25/5c7a1693-62f2-454e-9406-0b4a132ebf25.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.965294] env[65788]: DEBUG nova.network.neutron [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Successfully created port: a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 778.967485] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91725d9e-ae8a-4f55-b723-c1c634566eb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.989543] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 778.989543] env[65788]: value = "task-4662275" [ 778.989543] env[65788]: _type = "Task" [ 778.989543] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.002164] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662275, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.040841] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 779.184160] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 779.184160] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 779.184160] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 779.184160] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 779.184373] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 779.184373] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 779.185067] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.185594] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 779.186529] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 779.186529] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 779.188155] env[65788]: DEBUG nova.virt.hardware [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 779.192812] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Reconfiguring VM instance instance-00000009 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 779.195923] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11692878-392b-43c6-a8b3-0a337a8d7fae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.217434] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 779.217434] env[65788]: value = "task-4662276" [ 779.217434] env[65788]: _type = "Task" [ 779.217434] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.234705] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662276, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.262458] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662274, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.286331] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5243eb71-f25d-8fcb-c9bf-4a3386c7b214, 'name': SearchDatastore_Task, 'duration_secs': 0.059057} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.286912] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.287267] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.287716] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-865c598e-6f5f-4876-ac05-ce703735a76a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.299225] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 779.299225] env[65788]: value = "task-4662277" [ 779.299225] env[65788]: _type = "Task" [ 779.299225] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.309060] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.503669] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662275, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.630447] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3275ed-eb75-47d7-ab44-d74652e47550 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.638934] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523d717f-b1b8-4345-8381-0e2fb7f3b668 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.674216] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefec6d2-adde-4d10-a458-8f0aa6fae3da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.682862] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf42303-557b-4c30-b417-928e928426d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.697957] env[65788]: DEBUG nova.compute.provider_tree [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.728202] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662276, 'name': ReconfigVM_Task, 'duration_secs': 0.394258} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.728350] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Reconfigured VM instance instance-00000009 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 779.729146] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd4cf7a-bd9c-4437-9fa3-12a4b929d829 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.752636] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa/eceafff8-7d28-4b9b-ade6-5736d5977efa.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.753015] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bea5500-8ab3-40ab-bb8b-2ef5d78bd1d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.775168] env[65788]: DEBUG oslo_vmware.api [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662274, 'name': PowerOnVM_Task, 'duration_secs': 0.521328} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.776455] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 779.778238] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 779.778238] env[65788]: value = "task-4662278" [ 779.778238] env[65788]: _type = "Task" [ 779.778238] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.779583] env[65788]: DEBUG nova.compute.manager [None req-ad153bdf-d723-405e-97ef-1ab8117824be tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 779.783339] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146e0889-e1be-4676-8aaa-3d5560723462 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.796899] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662278, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.810036] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662277, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.892276] env[65788]: DEBUG nova.compute.manager [req-ef9c26ab-b8e1-4799-9f19-1956765c06bc req-721bd127-4b21-4ce7-993c-506ea28ccab5 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Received event network-vif-deleted-f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 779.892276] env[65788]: INFO nova.compute.manager [req-ef9c26ab-b8e1-4799-9f19-1956765c06bc req-721bd127-4b21-4ce7-993c-506ea28ccab5 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Neutron deleted interface f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8; detaching it from the instance and deleting it from the info cache [ 779.893877] env[65788]: DEBUG nova.network.neutron [req-ef9c26ab-b8e1-4799-9f19-1956765c06bc req-721bd127-4b21-4ce7-993c-506ea28ccab5 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 779.914772] env[65788]: DEBUG nova.network.neutron [-] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.010738] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662275, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.055778] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 780.088603] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 780.088874] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.089076] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 780.089823] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.089823] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 780.089823] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 780.089948] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.090153] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 780.090428] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 780.090754] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 780.091365] env[65788]: DEBUG nova.virt.hardware [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 780.092436] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ceef7b-d084-4f0e-aaee-ad8c2bb33241 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.104759] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2968ce7-335d-4ba7-9859-1d19fe6e31ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.202494] env[65788]: DEBUG nova.scheduler.client.report [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.289922] env[65788]: DEBUG oslo_vmware.api [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662278, 'name': ReconfigVM_Task, 'duration_secs': 0.487031} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.290425] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Reconfigured VM instance instance-00000009 to attach disk [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa/eceafff8-7d28-4b9b-ade6-5736d5977efa.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.290930] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 780.317101] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.940228} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.317979] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.318356] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.318799] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b967d236-0db8-4036-8336-f435de769505 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.329319] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 780.329319] env[65788]: value = "task-4662279" [ 780.329319] env[65788]: _type = "Task" [ 780.329319] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.342104] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662279, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.399682] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c524c88-9f2e-48c2-adcd-cbf952c1571d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.413353] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aaae413-7af7-46e5-8af0-8044a9278d75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.425532] env[65788]: INFO nova.compute.manager [-] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Took 1.58 seconds to deallocate network for instance. [ 780.432391] env[65788]: DEBUG nova.compute.manager [req-1c1a95b0-faa1-40c6-9e2b-0138205c0ba2 req-fac0b796-03d2-4328-83bb-811f4f5abf34 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received event network-vif-plugged-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 780.432618] env[65788]: DEBUG oslo_concurrency.lockutils [req-1c1a95b0-faa1-40c6-9e2b-0138205c0ba2 req-fac0b796-03d2-4328-83bb-811f4f5abf34 service nova] Acquiring lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.432812] env[65788]: DEBUG oslo_concurrency.lockutils [req-1c1a95b0-faa1-40c6-9e2b-0138205c0ba2 req-fac0b796-03d2-4328-83bb-811f4f5abf34 service nova] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.432976] env[65788]: DEBUG oslo_concurrency.lockutils [req-1c1a95b0-faa1-40c6-9e2b-0138205c0ba2 req-fac0b796-03d2-4328-83bb-811f4f5abf34 service nova] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.433158] env[65788]: DEBUG nova.compute.manager [req-1c1a95b0-faa1-40c6-9e2b-0138205c0ba2 req-fac0b796-03d2-4328-83bb-811f4f5abf34 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] No waiting events found dispatching network-vif-plugged-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 780.433319] env[65788]: WARNING nova.compute.manager [req-1c1a95b0-faa1-40c6-9e2b-0138205c0ba2 req-fac0b796-03d2-4328-83bb-811f4f5abf34 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received unexpected event network-vif-plugged-a52878da-edb4-4e48-8028-2ef5a48d8253 for instance with vm_state building and task_state spawning. [ 780.453031] env[65788]: DEBUG nova.compute.manager [req-ef9c26ab-b8e1-4799-9f19-1956765c06bc req-721bd127-4b21-4ce7-993c-506ea28ccab5 service nova] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Detach interface failed, port_id=f24d1ba6-c026-4b7b-ab0c-ad67cb1fc5f8, reason: Instance 670f59c3-10b0-431f-a8b0-ef6f82a938be could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 780.503663] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662275, 'name': ReconfigVM_Task, 'duration_secs': 1.313547} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.503954] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 5c7a1693-62f2-454e-9406-0b4a132ebf25/5c7a1693-62f2-454e-9406-0b4a132ebf25.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.504610] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37487693-8117-467b-a758-08a9f21b7980 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.509586] env[65788]: DEBUG nova.network.neutron [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Successfully updated port: a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 780.512646] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 780.512646] env[65788]: value = "task-4662280" [ 780.512646] env[65788]: _type = "Task" [ 780.512646] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.526228] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662280, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.708055] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.680s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.709099] env[65788]: DEBUG nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 780.711829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.554s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.714184] env[65788]: INFO nova.compute.claims [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.798634] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b5d1e1-467d-4d9c-a16e-08ec8c2e6c69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.820586] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ce9491-a59b-49e0-b2ad-7795e4113d0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.840283] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 780.852605] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662279, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070348} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.852866] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 780.853771] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ede900e-fc38-49b2-8e12-64fba6c75aa3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.882667] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.883259] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dee74f7-56bf-4999-96e0-287d4ebf19df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.909400] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 780.909400] env[65788]: value = "task-4662281" [ 780.909400] env[65788]: _type = "Task" [ 780.909400] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.918712] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662281, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.936183] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 781.014155] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.014709] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.014709] env[65788]: DEBUG nova.network.neutron [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 781.028858] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662280, 'name': Rename_Task, 'duration_secs': 0.14859} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.029778] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 781.030013] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1fce9a4-4612-4403-b076-70025749219b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.040217] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 781.040217] env[65788]: value = "task-4662282" [ 781.040217] env[65788]: _type = "Task" [ 781.040217] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.056017] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.223616] env[65788]: DEBUG nova.compute.utils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 781.225720] env[65788]: DEBUG nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 781.348735] env[65788]: WARNING neutronclient.v2_0.client [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.388174] env[65788]: DEBUG nova.network.neutron [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Port d6e04a22-9f86-43c9-ac42-95191eaf454b binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 781.421851] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662281, 'name': ReconfigVM_Task, 'duration_secs': 0.362382} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.422461] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Reconfigured VM instance instance-00000010 to attach disk [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904/5cef20ed-d9f0-4237-ae9d-da401b953904.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.423425] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24b80931-01b9-4037-9be0-e91d8ae2da34 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.432447] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 781.432447] env[65788]: value = "task-4662283" [ 781.432447] env[65788]: _type = "Task" [ 781.432447] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.442047] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662283, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.521865] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.521865] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.553934] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662282, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.563047] env[65788]: DEBUG nova.network.neutron [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 781.582915] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.583331] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.645912] env[65788]: WARNING neutronclient.v2_0.client [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 781.646614] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.646945] env[65788]: WARNING openstack [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.729824] env[65788]: DEBUG nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 781.739834] env[65788]: DEBUG nova.network.neutron [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 781.946701] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662283, 'name': Rename_Task, 'duration_secs': 0.255944} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.947101] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 781.947251] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d351220f-e52d-439c-808a-a10adb16b9b0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.955066] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 781.955066] env[65788]: value = "task-4662284" [ 781.955066] env[65788]: _type = "Task" [ 781.955066] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.963926] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662284, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.054381] env[65788]: DEBUG oslo_vmware.api [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662282, 'name': PowerOnVM_Task, 'duration_secs': 0.530482} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.054677] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.054860] env[65788]: INFO nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Took 9.20 seconds to spawn the instance on the hypervisor. [ 782.055071] env[65788]: DEBUG nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 782.055895] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f6a4a7-08c7-42e3-a1e8-71ecedf7b15e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.244425] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.244425] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Instance network_info: |[{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 782.244697] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:e9:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c791d09c-1086-4ee1-bcde-6ca7d259cabd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a52878da-edb4-4e48-8028-2ef5a48d8253', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.252972] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Creating folder: Project (41f69a1056bc454e83bd9561cf12c506). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.256559] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1c269c0-1b3a-440f-bed2-b9d80dc45164 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.271173] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Created folder: Project (41f69a1056bc454e83bd9561cf12c506) in parent group-v910111. [ 782.271173] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Creating folder: Instances. Parent ref: group-v910259. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.271173] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9736a99b-19fa-4aca-8b39-126b12ae4529 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.274767] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d35c3c3-e9df-4383-bc70-2e2e199c9911 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.284329] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f49421-b9f0-4676-9f97-14111e286976 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.289014] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Created folder: Instances in parent group-v910259. [ 782.289327] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 782.289939] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.290244] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-387dd5ce-a96d-4b44-99f6-98c40597aaaa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.335946] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f459d060-6552-4bd5-936b-830e664d14ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.341433] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.341433] env[65788]: value = "task-4662287" [ 782.341433] env[65788]: _type = "Task" [ 782.341433] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.347726] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016dc4de-6338-4566-8b49-08965a578dbf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.358862] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662287, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.367674] env[65788]: DEBUG nova.compute.provider_tree [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.415870] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 782.416124] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 782.416301] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.466567] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662284, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.480961] env[65788]: DEBUG nova.compute.manager [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 782.481175] env[65788]: DEBUG nova.compute.manager [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing instance network info cache due to event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 782.481434] env[65788]: DEBUG oslo_concurrency.lockutils [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.481679] env[65788]: DEBUG oslo_concurrency.lockutils [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.481754] env[65788]: DEBUG nova.network.neutron [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 782.578598] env[65788]: INFO nova.compute.manager [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Took 39.12 seconds to build instance. [ 782.727086] env[65788]: INFO nova.compute.manager [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Rescuing [ 782.727437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.727746] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.728030] env[65788]: DEBUG nova.network.neutron [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 782.739756] env[65788]: DEBUG nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 782.769838] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 782.770143] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 782.770303] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 782.770491] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 782.770641] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 782.770785] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 782.770993] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.771171] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 782.771377] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 782.771548] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 782.771720] env[65788]: DEBUG nova.virt.hardware [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 782.772681] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cafa241-16a3-4d1f-8e7d-cd41f7734de8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.783907] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16be9a96-b426-4427-a5d4-6230de4bc8fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.801263] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.807239] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Creating folder: Project (d4b430072df249edba814579983d466d). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.807590] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b9c42449-67a5-434f-a6ca-f7eb01ac7fb6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.823544] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Created folder: Project (d4b430072df249edba814579983d466d) in parent group-v910111. [ 782.823811] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Creating folder: Instances. Parent ref: group-v910262. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.824213] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbe6c80c-a21f-433a-8887-09456f2a7ca8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.836237] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Created folder: Instances in parent group-v910262. [ 782.836652] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 782.836890] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.837138] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6ab3a3a-c426-4ec7-b57a-c87e968a6a17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.861453] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.861453] env[65788]: value = "task-4662290" [ 782.861453] env[65788]: _type = "Task" [ 782.861453] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.864670] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662287, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.876902] env[65788]: DEBUG nova.scheduler.client.report [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 782.881321] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662290, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.966430] env[65788]: DEBUG oslo_vmware.api [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662284, 'name': PowerOnVM_Task, 'duration_secs': 0.917827} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.966719] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.966925] env[65788]: DEBUG nova.compute.manager [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 782.967736] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2772d580-25a2-41c2-a3a6-c11b159c60a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.984659] env[65788]: WARNING neutronclient.v2_0.client [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 782.985301] env[65788]: WARNING openstack [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.985301] env[65788]: WARNING openstack [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.080936] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c939509b-57e7-4e7e-80d7-40ea918b758f tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.581s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.105810] env[65788]: WARNING openstack [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.106268] env[65788]: WARNING openstack [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.184508] env[65788]: WARNING neutronclient.v2_0.client [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.185618] env[65788]: WARNING openstack [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.186235] env[65788]: WARNING openstack [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.232080] env[65788]: WARNING neutronclient.v2_0.client [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.232904] env[65788]: WARNING openstack [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.233332] env[65788]: WARNING openstack [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.327535] env[65788]: DEBUG nova.network.neutron [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updated VIF entry in instance network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 783.327535] env[65788]: DEBUG nova.network.neutron [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.371479] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662287, 'name': CreateVM_Task, 'duration_secs': 0.540388} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.371970] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 783.372530] env[65788]: WARNING neutronclient.v2_0.client [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.372893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.373049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.373357] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 783.373629] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52358af8-573f-4637-badc-a57368b969af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.378163] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662290, 'name': CreateVM_Task, 'duration_secs': 0.467874} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.379408] env[65788]: WARNING openstack [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.379760] env[65788]: WARNING openstack [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.386378] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 783.386781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.675s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.387252] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 783.391171] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.391443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.391s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.391658] env[65788]: DEBUG nova.objects.instance [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lazy-loading 'resources' on Instance uuid f604c16d-2a86-40d3-9891-5b33309b3047 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 783.392974] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 783.392974] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bad615-e32c-a697-4d70-c5af7f9daa3d" [ 783.392974] env[65788]: _type = "Task" [ 783.392974] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.406029] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bad615-e32c-a697-4d70-c5af7f9daa3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.423071] env[65788]: WARNING neutronclient.v2_0.client [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.459566] env[65788]: WARNING neutronclient.v2_0.client [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.460342] env[65788]: WARNING openstack [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.460594] env[65788]: WARNING openstack [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.471735] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.471893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.472082] env[65788]: DEBUG nova.network.neutron [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 783.487268] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.544455] env[65788]: DEBUG nova.network.neutron [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Updating instance_info_cache with network_info: [{"id": "f116c9d8-4c98-4af9-8747-0c585d99d738", "address": "fa:16:3e:e3:a3:1e", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf116c9d8-4c", "ovs_interfaceid": "f116c9d8-4c98-4af9-8747-0c585d99d738", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.656547] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a73951-18af-4819-a3e3-c65ebfc3741c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.664917] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Suspending the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 783.665212] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-db67823b-2fb3-4632-98d1-bba42fe6d340 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.673296] env[65788]: DEBUG oslo_vmware.api [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] Waiting for the task: (returnval){ [ 783.673296] env[65788]: value = "task-4662291" [ 783.673296] env[65788]: _type = "Task" [ 783.673296] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.683635] env[65788]: DEBUG oslo_vmware.api [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] Task: {'id': task-4662291, 'name': SuspendVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.827915] env[65788]: DEBUG oslo_concurrency.lockutils [req-cbafa766-fe94-42a6-9e92-5345ac10c1c9 req-bf3fdfd9-7230-40b0-94e4-1b7a02de062a service nova] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.896044] env[65788]: DEBUG nova.compute.utils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 783.903965] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 783.903965] env[65788]: DEBUG nova.network.neutron [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 783.903965] env[65788]: WARNING neutronclient.v2_0.client [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.903965] env[65788]: WARNING neutronclient.v2_0.client [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.905159] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.905650] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.926271] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bad615-e32c-a697-4d70-c5af7f9daa3d, 'name': SearchDatastore_Task, 'duration_secs': 0.012143} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.930029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.930218] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.930546] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.930760] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.931117] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.934549] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.934916] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 783.935301] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f513205-3baa-4bf9-94d9-fafc1b0e2614 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.938552] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b860baa-19d0-4f90-8517-4685ea322587 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.950134] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 783.950134] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f8731f-0237-178f-48fa-120a96dc655e" [ 783.950134] env[65788]: _type = "Task" [ 783.950134] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.950464] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.950674] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.954766] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-798c22df-e1c3-4d91-8ac6-103773b677c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.963913] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 783.963913] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e35b2-b678-5eee-782d-bb988c713b01" [ 783.963913] env[65788]: _type = "Task" [ 783.963913] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.973176] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f8731f-0237-178f-48fa-120a96dc655e, 'name': SearchDatastore_Task, 'duration_secs': 0.015405} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.977808] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.978075] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.978464] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.978779] env[65788]: WARNING neutronclient.v2_0.client [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 783.979435] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.979782] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.997562] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e35b2-b678-5eee-782d-bb988c713b01, 'name': SearchDatastore_Task, 'duration_secs': 0.012659} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.001928] env[65788]: DEBUG nova.policy [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61663f00d71a465ab1ec9db56880da0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fb2fb1250354a1ba3e151b909c67667', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 784.003976] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0cd4de2-8460-4627-a378-325b545228f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.014272] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 784.014272] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bd8f75-fadd-f626-55d1-af44bc92f716" [ 784.014272] env[65788]: _type = "Task" [ 784.014272] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.028744] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bd8f75-fadd-f626-55d1-af44bc92f716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.048138] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.140285] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 784.140697] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 784.188922] env[65788]: DEBUG oslo_vmware.api [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] Task: {'id': task-4662291, 'name': SuspendVM_Task} progress is 62%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.236112] env[65788]: WARNING neutronclient.v2_0.client [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 784.237428] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 784.237428] env[65788]: WARNING openstack [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 784.333656] env[65788]: DEBUG nova.network.neutron [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 784.403430] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 784.421218] env[65788]: DEBUG nova.network.neutron [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Successfully created port: 05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 784.527163] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bd8f75-fadd-f626-55d1-af44bc92f716, 'name': SearchDatastore_Task, 'duration_secs': 0.014073} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.528027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.528027] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9/2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.528027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 784.528287] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.528393] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c23c6c1-8251-41f2-baba-d98c6a274784 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.530684] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31f3a576-124a-4e31-ad92-a039801f8347 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.541318] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 784.541318] env[65788]: value = "task-4662292" [ 784.541318] env[65788]: _type = "Task" [ 784.541318] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.544806] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.544806] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 784.545699] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c463719-85e8-4247-bb98-0aa3c31cd7a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.551218] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662292, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.556250] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 784.556250] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520065c6-2dbc-868e-edae-d3900f9c957f" [ 784.556250] env[65788]: _type = "Task" [ 784.556250] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.558123] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523fe100-543c-48a8-be08-64183938a948 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.570320] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520065c6-2dbc-868e-edae-d3900f9c957f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.572431] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d93fae5-62bc-43b7-8f02-d461027a30ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.612337] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a47fe4-93a3-4d8c-aa80-f08e8af7c504 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.629453] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e96b93-6d0a-4ac5-a581-ed5e123b1d71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.648056] env[65788]: DEBUG nova.compute.provider_tree [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.687363] env[65788]: DEBUG oslo_vmware.api [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] Task: {'id': task-4662291, 'name': SuspendVM_Task, 'duration_secs': 0.752259} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.687651] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Suspended the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 784.687844] env[65788]: DEBUG nova.compute.manager [None req-d0de81b4-45c1-49dd-b981-b8ba6d8497bc tempest-ServersAdminNegativeTestJSON-1486293406 tempest-ServersAdminNegativeTestJSON-1486293406-project-admin] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 784.688667] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ad354d-196f-4e44-b934-e90b2754234e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.837188] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 785.051531] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662292, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.071396] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520065c6-2dbc-868e-edae-d3900f9c957f, 'name': SearchDatastore_Task, 'duration_secs': 0.03626} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.072951] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcedb25d-99e1-46db-9608-28167ef2a0f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.081429] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 785.081429] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5207b1c5-69a6-911d-c305-dd96e41331ca" [ 785.081429] env[65788]: _type = "Task" [ 785.081429] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.095801] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5207b1c5-69a6-911d-c305-dd96e41331ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.123076] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "f3c53f67-85bf-4c18-9313-75eb90862f78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.123214] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.123419] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "f3c53f67-85bf-4c18-9313-75eb90862f78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.123949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.123949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.126475] env[65788]: INFO nova.compute.manager [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Terminating instance [ 785.151125] env[65788]: DEBUG nova.scheduler.client.report [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 785.366878] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b795f9-5050-4c14-8646-bbbeaeaf5994 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.387870] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2924cf-cc12-43ee-81a2-eb66d1d9117c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.396619] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 785.414234] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 785.442103] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 785.442376] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 785.442535] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 785.442716] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 785.442859] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 785.443010] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 785.443232] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 785.443388] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 785.443573] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 785.443763] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 785.443937] env[65788]: DEBUG nova.virt.hardware [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 785.444827] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0947bc43-2078-46c3-bee3-d464680ffc9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.455255] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e5a671-88f5-458a-b629-a0a5a61e09ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.551338] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662292, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.797606} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.551640] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9/2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.551851] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.552127] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61ce99d2-a581-4f22-8eed-95cbc8e952ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.559656] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 785.559656] env[65788]: value = "task-4662293" [ 785.559656] env[65788]: _type = "Task" [ 785.559656] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.571060] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.594811] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5207b1c5-69a6-911d-c305-dd96e41331ca, 'name': SearchDatastore_Task, 'duration_secs': 0.073928} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.595185] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 785.595516] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 785.595859] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5054c4c5-4b66-41b2-9886-b7e20623356a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.604380] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 785.604380] env[65788]: value = "task-4662294" [ 785.604380] env[65788]: _type = "Task" [ 785.604380] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.614322] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662294, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.615948] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.616239] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a55886c7-cd6d-4625-86fd-dfbf8cc1f4b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.623016] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 785.623016] env[65788]: value = "task-4662295" [ 785.623016] env[65788]: _type = "Task" [ 785.623016] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.632487] env[65788]: DEBUG nova.compute.manager [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 785.632750] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.633100] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662295, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.634224] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a49d8bc-30a7-4acd-ba2c-bfacba109b80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.642138] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.642395] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dda10bfb-75a1-4a05-97a6-5ac91fb14e5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.650354] env[65788]: DEBUG oslo_vmware.api [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 785.650354] env[65788]: value = "task-4662296" [ 785.650354] env[65788]: _type = "Task" [ 785.650354] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.658248] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.265s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.659720] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.584s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.661993] env[65788]: INFO nova.compute.claims [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.672150] env[65788]: DEBUG oslo_vmware.api [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.684749] env[65788]: INFO nova.scheduler.client.report [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Deleted allocations for instance f604c16d-2a86-40d3-9891-5b33309b3047 [ 785.905826] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4bfcf54c-0d7a-48a3-8d6a-385b09ff821b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance 'eceafff8-7d28-4b9b-ade6-5736d5977efa' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 786.067782] env[65788]: DEBUG nova.network.neutron [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Successfully updated port: 05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 786.072411] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080564} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.072986] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.074032] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca74e5d-f3ac-46f6-a857-848d235966ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.101431] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9/2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.102480] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6106b50-aa70-4048-9465-e6a1ea6b54fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.131482] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 786.131482] env[65788]: value = "task-4662297" [ 786.131482] env[65788]: _type = "Task" [ 786.131482] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.131738] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662294, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.138829] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662295, 'name': PowerOffVM_Task, 'duration_secs': 0.270802} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.139534] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.141084] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c33ab7b-0d2f-42fd-9533-a07cfcabdf1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.149624] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662297, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.151454] env[65788]: DEBUG nova.compute.manager [req-b082ed3e-50b8-4c36-be84-26f76214ee23 req-d32b0ca2-a688-4b01-a639-d395cf4e6b4a service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Received event network-vif-plugged-05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 786.151745] env[65788]: DEBUG oslo_concurrency.lockutils [req-b082ed3e-50b8-4c36-be84-26f76214ee23 req-d32b0ca2-a688-4b01-a639-d395cf4e6b4a service nova] Acquiring lock "2c24bc30-d413-4714-81c2-b657cafe94bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.151973] env[65788]: DEBUG oslo_concurrency.lockutils [req-b082ed3e-50b8-4c36-be84-26f76214ee23 req-d32b0ca2-a688-4b01-a639-d395cf4e6b4a service nova] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.152260] env[65788]: DEBUG oslo_concurrency.lockutils [req-b082ed3e-50b8-4c36-be84-26f76214ee23 req-d32b0ca2-a688-4b01-a639-d395cf4e6b4a service nova] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.152475] env[65788]: DEBUG nova.compute.manager [req-b082ed3e-50b8-4c36-be84-26f76214ee23 req-d32b0ca2-a688-4b01-a639-d395cf4e6b4a service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] No waiting events found dispatching network-vif-plugged-05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 786.152672] env[65788]: WARNING nova.compute.manager [req-b082ed3e-50b8-4c36-be84-26f76214ee23 req-d32b0ca2-a688-4b01-a639-d395cf4e6b4a service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Received unexpected event network-vif-plugged-05d3722d-6c89-4066-84f7-83e285df6855 for instance with vm_state building and task_state spawning. [ 786.176350] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71cf236-7524-4523-956d-11a6c3840cd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.187024] env[65788]: DEBUG oslo_vmware.api [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662296, 'name': PowerOffVM_Task, 'duration_secs': 0.259641} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.187024] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.187024] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 786.187024] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-158f773c-03dd-4e47-b71c-281b2f8c8296 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.195365] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a5a9a5f-de38-48a9-96f7-77123f2a0bb3 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.757s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.196617] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 32.565s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.197298] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 786.197590] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.197712] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.200768] env[65788]: INFO nova.compute.manager [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Terminating instance [ 786.216738] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.216886] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7992b5d-cd0d-4362-97cf-a7502b936564 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.224407] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 786.224407] env[65788]: value = "task-4662299" [ 786.224407] env[65788]: _type = "Task" [ 786.224407] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.237192] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 786.237424] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.237645] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.237789] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.237952] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.238248] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e7b9d97-2a1d-40d4-a925-4186412370e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.250137] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.250355] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 786.251140] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ace9906d-ce46-4efa-acc3-f86b2586bf44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.257349] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 786.257349] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be4796-654e-1ae9-f782-6d88a0a5f481" [ 786.257349] env[65788]: _type = "Task" [ 786.257349] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.267696] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be4796-654e-1ae9-f782-6d88a0a5f481, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.282029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 786.282285] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 786.282500] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleting the datastore file [datastore2] f3c53f67-85bf-4c18-9313-75eb90862f78 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 786.282771] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7498698f-414b-4e5d-8b22-97b2b8f0da8b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.289976] env[65788]: DEBUG oslo_vmware.api [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 786.289976] env[65788]: value = "task-4662300" [ 786.289976] env[65788]: _type = "Task" [ 786.289976] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.298881] env[65788]: DEBUG oslo_vmware.api [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.573918] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "refresh_cache-2c24bc30-d413-4714-81c2-b657cafe94bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.574031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired lock "refresh_cache-2c24bc30-d413-4714-81c2-b657cafe94bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.574200] env[65788]: DEBUG nova.network.neutron [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 786.628440] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662294, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622015} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.628718] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 786.629081] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 786.629193] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e18fa58-979b-4bde-b76d-7bfd0babbc6e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.636925] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 786.636925] env[65788]: value = "task-4662301" [ 786.636925] env[65788]: _type = "Task" [ 786.636925] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.643306] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662297, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.649794] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.705744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.705744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquired lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.705898] env[65788]: DEBUG nova.network.neutron [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 786.772054] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be4796-654e-1ae9-f782-6d88a0a5f481, 'name': SearchDatastore_Task, 'duration_secs': 0.021974} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.773031] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ee3d8cc-0d6a-41b7-a124-449c24c47152 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.780710] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 786.780710] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5214ebec-e7f6-e903-00ea-bec0b2b4feca" [ 786.780710] env[65788]: _type = "Task" [ 786.780710] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.795082] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5214ebec-e7f6-e903-00ea-bec0b2b4feca, 'name': SearchDatastore_Task, 'duration_secs': 0.010582} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.799744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.799995] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. {{(pid=65788) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 786.800471] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-591050ec-38a0-4b34-a63e-7636a5817a04 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.815634] env[65788]: DEBUG oslo_vmware.api [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.486762} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.817260] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.817477] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 786.817674] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.817862] env[65788]: INFO nova.compute.manager [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Took 1.19 seconds to destroy the instance on the hypervisor. [ 786.818249] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 786.818514] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 786.818514] env[65788]: value = "task-4662302" [ 786.818514] env[65788]: _type = "Task" [ 786.818514] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.822087] env[65788]: DEBUG nova.compute.manager [-] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 786.822087] env[65788]: DEBUG nova.network.neutron [-] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 786.822264] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.822809] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.823090] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.841245] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.873810] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.078194] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.078831] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.130983] env[65788]: DEBUG nova.network.neutron [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 787.153090] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662297, 'name': ReconfigVM_Task, 'duration_secs': 1.005011} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.158707] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9/2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.159459] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158816} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.161175] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df2b86dd-a322-45a7-9b89-8139a4730208 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.163179] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 787.164011] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e81785-cfe5-429d-9cc4-a9c1bbc4979c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.188303] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 787.190776] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.191765] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.199530] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbb1cca9-e71b-48cf-aaa0-fa14c564c9df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.215145] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 787.215145] env[65788]: value = "task-4662303" [ 787.215145] env[65788]: _type = "Task" [ 787.215145] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.218538] env[65788]: DEBUG nova.compute.utils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Can not refresh info_cache because instance was not found {{(pid=65788) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 787.218974] env[65788]: WARNING neutronclient.v2_0.client [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.219644] env[65788]: WARNING openstack [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.220067] env[65788]: WARNING openstack [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.234709] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 787.234709] env[65788]: value = "task-4662304" [ 787.234709] env[65788]: _type = "Task" [ 787.234709] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.246297] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662303, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.257938] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662304, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.261703] env[65788]: DEBUG nova.network.neutron [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 787.309438] env[65788]: WARNING neutronclient.v2_0.client [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.310539] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.310715] env[65788]: WARNING openstack [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.338814] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fef3c5-8ebf-4ff0-9586-c03f75e4c6a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.356275] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662302, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.362282] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0245c9-482d-4ba7-a571-136515ff05e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.408063] env[65788]: DEBUG nova.network.neutron [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.410136] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6cd4ff-060d-4b7d-8a78-45845838a039 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.422173] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4133abbc-2265-466e-a3c3-aec448075af7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.451845] env[65788]: DEBUG nova.compute.provider_tree [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.456651] env[65788]: DEBUG nova.network.neutron [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Updating instance_info_cache with network_info: [{"id": "05d3722d-6c89-4066-84f7-83e285df6855", "address": "fa:16:3e:0c:13:2d", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05d3722d-6c", "ovs_interfaceid": "05d3722d-6c89-4066-84f7-83e285df6855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.598803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.599555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.599703] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.600455] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.600455] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.605157] env[65788]: INFO nova.compute.manager [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Terminating instance [ 787.668489] env[65788]: DEBUG nova.network.neutron [-] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.742662] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662303, 'name': Rename_Task, 'duration_secs': 0.221295} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.743787] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.743787] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b366ce98-ec3e-4474-923c-a64f53ea9fd9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.748882] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662304, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.750205] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 787.750205] env[65788]: value = "task-4662305" [ 787.750205] env[65788]: _type = "Task" [ 787.750205] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.758796] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662305, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.842517] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.846188} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.842895] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. [ 787.844046] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44911a26-ca06-41d6-9a2e-723475d2457f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.875750] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 787.876506] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1390cff-5f76-48a2-8c1a-09d07deb0b3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.898729] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 787.898729] env[65788]: value = "task-4662306" [ 787.898729] env[65788]: _type = "Task" [ 787.898729] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.909708] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662306, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.915919] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Releasing lock "refresh_cache-f604c16d-2a86-40d3-9891-5b33309b3047" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.915919] env[65788]: DEBUG nova.compute.manager [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 787.916221] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.916578] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7578abc-fc8f-486f-824e-d0241b8dbe68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.927498] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257a9115-84f6-4d56-b65c-03ebf29ff6e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.968935] env[65788]: DEBUG nova.scheduler.client.report [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 787.972099] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Releasing lock "refresh_cache-2c24bc30-d413-4714-81c2-b657cafe94bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.972482] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Instance network_info: |[{"id": "05d3722d-6c89-4066-84f7-83e285df6855", "address": "fa:16:3e:0c:13:2d", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05d3722d-6c", "ovs_interfaceid": "05d3722d-6c89-4066-84f7-83e285df6855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 787.973081] env[65788]: WARNING nova.virt.vmwareapi.vmops [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f604c16d-2a86-40d3-9891-5b33309b3047 could not be found. [ 787.973296] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.973424] env[65788]: INFO nova.compute.manager [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Took 0.06 seconds to destroy the instance on the hypervisor. [ 787.973696] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 787.974366] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:13:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05d3722d-6c89-4066-84f7-83e285df6855', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.981851] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating folder: Project (1fb2fb1250354a1ba3e151b909c67667). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.982138] env[65788]: DEBUG nova.compute.manager [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 787.982247] env[65788]: DEBUG nova.network.neutron [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 787.982513] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 787.983052] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 787.983315] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 787.990524] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1efaa850-f033-434e-a92d-b08331cba9ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.006027] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Created folder: Project (1fb2fb1250354a1ba3e151b909c67667) in parent group-v910111. [ 788.006027] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating folder: Instances. Parent ref: group-v910265. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.006027] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6ddd88b-4f00-4bdc-af4b-9bfce2a91fba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.007359] env[65788]: DEBUG nova.network.neutron [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 788.007639] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.016926] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Created folder: Instances in parent group-v910265. [ 788.017286] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 788.018172] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.018441] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6bac972a-7c3d-4042-9e4f-b41de6a68102 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.043233] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.043233] env[65788]: value = "task-4662309" [ 788.043233] env[65788]: _type = "Task" [ 788.043233] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.053721] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662309, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.110816] env[65788]: DEBUG nova.compute.manager [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 788.110816] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.111485] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebbb3641-0720-4593-98ff-54b0f78108cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.121122] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.121584] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac5a8f53-02db-45ed-b9ba-27838ff70e66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.130732] env[65788]: DEBUG oslo_vmware.api [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 788.130732] env[65788]: value = "task-4662310" [ 788.130732] env[65788]: _type = "Task" [ 788.130732] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.142993] env[65788]: DEBUG oslo_vmware.api [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.171962] env[65788]: INFO nova.compute.manager [-] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Took 1.35 seconds to deallocate network for instance. [ 788.214871] env[65788]: DEBUG nova.compute.manager [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Received event network-changed-05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 788.214871] env[65788]: DEBUG nova.compute.manager [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Refreshing instance network info cache due to event network-changed-05d3722d-6c89-4066-84f7-83e285df6855. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 788.215087] env[65788]: DEBUG oslo_concurrency.lockutils [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Acquiring lock "refresh_cache-2c24bc30-d413-4714-81c2-b657cafe94bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.216359] env[65788]: DEBUG oslo_concurrency.lockutils [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Acquired lock "refresh_cache-2c24bc30-d413-4714-81c2-b657cafe94bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.216359] env[65788]: DEBUG nova.network.neutron [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Refreshing network info cache for port 05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 788.250616] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662304, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.260773] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662305, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.409943] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662306, 'name': ReconfigVM_Task, 'duration_secs': 0.442402} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.410269] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 788.411188] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b3e163-1991-4a41-9858-7f329fbe71cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.438960] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a2bc9ef-f26f-4caf-ad14-a44c6acfcaca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.455894] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 788.455894] env[65788]: value = "task-4662311" [ 788.455894] env[65788]: _type = "Task" [ 788.455894] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.465330] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662311, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.475473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.816s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.476073] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 788.478986] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.128s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.479232] env[65788]: DEBUG nova.objects.instance [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lazy-loading 'resources' on Instance uuid 72111766-217d-469f-ad92-ca0ce0e67090 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.509666] env[65788]: DEBUG nova.network.neutron [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 788.554393] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662309, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.558978] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "5c7a1693-62f2-454e-9406-0b4a132ebf25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.559249] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.559582] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "5c7a1693-62f2-454e-9406-0b4a132ebf25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.559655] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.559792] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.562089] env[65788]: INFO nova.compute.manager [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Terminating instance [ 788.641134] env[65788]: DEBUG oslo_vmware.api [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662310, 'name': PowerOffVM_Task, 'duration_secs': 0.24029} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.641798] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.642134] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 788.642846] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a8aa148-794a-422a-961e-f3d420821f58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.680098] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.719045] env[65788]: WARNING neutronclient.v2_0.client [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.719968] env[65788]: WARNING openstack [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.720373] env[65788]: WARNING openstack [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.729392] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 788.729588] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 788.729765] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Deleting the datastore file [datastore2] 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.730070] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a56c34d0-9e82-468b-a7ec-41eaa5275d75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.738040] env[65788]: DEBUG oslo_vmware.api [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for the task: (returnval){ [ 788.738040] env[65788]: value = "task-4662313" [ 788.738040] env[65788]: _type = "Task" [ 788.738040] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.749376] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662304, 'name': ReconfigVM_Task, 'duration_secs': 1.314004} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.752601] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Reconfigured VM instance instance-00000032 to attach disk [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 788.753452] env[65788]: DEBUG oslo_vmware.api [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.753690] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0cbc126-fa34-4ed5-b71d-b7b693c1daca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.770224] env[65788]: DEBUG oslo_vmware.api [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662305, 'name': PowerOnVM_Task, 'duration_secs': 0.549495} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.771643] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.771858] env[65788]: INFO nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Took 8.72 seconds to spawn the instance on the hypervisor. [ 788.772055] env[65788]: DEBUG nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 788.772402] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 788.772402] env[65788]: value = "task-4662314" [ 788.772402] env[65788]: _type = "Task" [ 788.772402] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.773129] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c21097-d6dd-4fec-9871-d51397f3ba44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.776198] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.776198] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.776369] env[65788]: DEBUG nova.compute.manager [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Going to confirm migration 2 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 788.792842] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662314, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.832036] env[65788]: WARNING openstack [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.832501] env[65788]: WARNING openstack [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.923960] env[65788]: WARNING neutronclient.v2_0.client [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.924686] env[65788]: WARNING openstack [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.925054] env[65788]: WARNING openstack [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.966893] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662311, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.984680] env[65788]: DEBUG nova.compute.utils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 788.989838] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 788.989838] env[65788]: DEBUG nova.network.neutron [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 788.990058] env[65788]: WARNING neutronclient.v2_0.client [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.990355] env[65788]: WARNING neutronclient.v2_0.client [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.990977] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.991331] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.013089] env[65788]: INFO nova.compute.manager [-] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Took 1.03 seconds to deallocate network for instance. [ 789.045099] env[65788]: DEBUG nova.network.neutron [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Updated VIF entry in instance network info cache for port 05d3722d-6c89-4066-84f7-83e285df6855. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 789.045573] env[65788]: DEBUG nova.network.neutron [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Updating instance_info_cache with network_info: [{"id": "05d3722d-6c89-4066-84f7-83e285df6855", "address": "fa:16:3e:0c:13:2d", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.92", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05d3722d-6c", "ovs_interfaceid": "05d3722d-6c89-4066-84f7-83e285df6855", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 789.058497] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662309, 'name': CreateVM_Task, 'duration_secs': 0.53434} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.058679] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.059223] env[65788]: WARNING neutronclient.v2_0.client [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.059723] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.059723] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.060013] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 789.060555] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54722996-d5a6-4b67-9136-e79c7505bb40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.068272] env[65788]: DEBUG nova.policy [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '61663f00d71a465ab1ec9db56880da0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fb2fb1250354a1ba3e151b909c67667', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 789.070370] env[65788]: DEBUG nova.compute.manager [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 789.070589] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.072476] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb33cdf-fba7-43d4-88ef-93a10a95cf0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.077361] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 789.077361] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52544c4d-7811-a438-bc6c-e7d09e951b98" [ 789.077361] env[65788]: _type = "Task" [ 789.077361] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.082530] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 789.088605] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28c8888e-f330-44b7-9ae2-0fade21da914 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.090471] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52544c4d-7811-a438-bc6c-e7d09e951b98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.171112] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 789.171621] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 789.171840] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Deleting the datastore file [datastore2] 5c7a1693-62f2-454e-9406-0b4a132ebf25 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.172148] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8aed6a5b-a99e-4bf0-81eb-18a6e34df5d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.183389] env[65788]: DEBUG oslo_vmware.api [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 789.183389] env[65788]: value = "task-4662316" [ 789.183389] env[65788]: _type = "Task" [ 789.183389] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.192333] env[65788]: DEBUG oslo_vmware.api [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.254340] env[65788]: DEBUG oslo_vmware.api [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Task: {'id': task-4662313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152172} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.254625] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.254811] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 789.255031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.255202] env[65788]: INFO nova.compute.manager [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 789.255497] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 789.255760] env[65788]: DEBUG nova.compute.manager [-] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 789.255829] env[65788]: DEBUG nova.network.neutron [-] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 789.256149] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.256763] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.257018] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.290523] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662314, 'name': Rename_Task, 'duration_secs': 0.193232} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.291765] env[65788]: WARNING neutronclient.v2_0.client [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.293587] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.294078] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d499b91-396e-41ec-be90-240a9f24eb87 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.308485] env[65788]: INFO nova.compute.manager [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Took 42.02 seconds to build instance. [ 789.310914] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 789.310914] env[65788]: value = "task-4662317" [ 789.310914] env[65788]: _type = "Task" [ 789.310914] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.322727] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.336560] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.399540] env[65788]: DEBUG nova.network.neutron [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Successfully created port: 4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 789.449770] env[65788]: WARNING neutronclient.v2_0.client [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.450129] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.450288] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.450455] env[65788]: DEBUG nova.network.neutron [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 789.450634] env[65788]: DEBUG nova.objects.instance [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'info_cache' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 789.470699] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662311, 'name': ReconfigVM_Task, 'duration_secs': 0.641496} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.471046] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.471411] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0131155a-f446-494d-b79d-86520a644293 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.480472] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 789.480472] env[65788]: value = "task-4662318" [ 789.480472] env[65788]: _type = "Task" [ 789.480472] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.492474] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 789.496493] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662318, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.521599] env[65788]: INFO nova.compute.manager [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance disappeared during terminate [ 789.521936] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e6189d4e-5884-41c6-80f0-7a955caa5346 tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "f604c16d-2a86-40d3-9891-5b33309b3047" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.325s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.548748] env[65788]: DEBUG oslo_concurrency.lockutils [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] Releasing lock "refresh_cache-2c24bc30-d413-4714-81c2-b657cafe94bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.548987] env[65788]: DEBUG nova.compute.manager [req-044b9e54-f7cf-42a3-b5cf-c0caa1329709 req-fa5a474f-edcb-4be8-8f5f-b249d9fc7767 service nova] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Received event network-vif-deleted-ea77fa15-d004-4f6d-bea0-02027383b034 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 789.564586] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca13143-0255-4509-bc8c-5bd8fb4e18ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.573455] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0e4edf-451b-446e-8345-e69429938ea4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.587700] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52544c4d-7811-a438-bc6c-e7d09e951b98, 'name': SearchDatastore_Task, 'duration_secs': 0.010715} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.614682] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.615149] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.615488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.615646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.615868] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.618216] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-398df215-1e86-45fc-a0f4-b00d8cd31466 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.621166] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcd013c-5f5a-447a-a7d5-7ff0f18b8ecf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.631867] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7d1bb4-ae45-4749-9d87-e34d6e345bab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.637422] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.637634] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 789.638703] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa39409e-24cc-4989-8198-11014e95e811 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.649862] env[65788]: DEBUG nova.compute.provider_tree [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.656833] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 789.656833] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520b4536-d787-de15-359e-dfffa01a0097" [ 789.656833] env[65788]: _type = "Task" [ 789.656833] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.670348] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520b4536-d787-de15-359e-dfffa01a0097, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.693335] env[65788]: DEBUG oslo_vmware.api [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.4908} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.693758] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.693957] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 789.694143] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.694313] env[65788]: INFO nova.compute.manager [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Took 0.62 seconds to destroy the instance on the hypervisor. [ 789.694562] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 789.694755] env[65788]: DEBUG nova.compute.manager [-] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 789.694848] env[65788]: DEBUG nova.network.neutron [-] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 789.695119] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.695760] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.696129] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.758270] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 789.812742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-77ee0ac6-0cf9-44ed-8dc1-f3ab5131e9c6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.737s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.824886] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662317, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.991652] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662318, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.125394] env[65788]: DEBUG nova.network.neutron [-] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.153323] env[65788]: DEBUG nova.scheduler.client.report [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 790.169718] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520b4536-d787-de15-359e-dfffa01a0097, 'name': SearchDatastore_Task, 'duration_secs': 0.022242} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.172821] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf530824-6e56-4521-ba47-252ba17e1514 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.178712] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 790.178712] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a6450f-99fa-2e63-38d9-6ad682b223bb" [ 790.178712] env[65788]: _type = "Task" [ 790.178712] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.188679] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a6450f-99fa-2e63-38d9-6ad682b223bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.206909] env[65788]: DEBUG nova.compute.manager [req-bcf9e7e4-f78a-4f6d-813b-1b9f2e62e522 req-95df9e58-a863-459c-b20e-3f4d9d75e994 service nova] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Received event network-vif-deleted-89c8bddf-8224-4440-87da-352af62539fc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 790.324995] env[65788]: DEBUG oslo_vmware.api [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662317, 'name': PowerOnVM_Task, 'duration_secs': 0.853903} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.325306] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.325509] env[65788]: INFO nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Took 7.59 seconds to spawn the instance on the hypervisor. [ 790.325689] env[65788]: DEBUG nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 790.326517] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8606169c-177e-4e4a-9959-74da3ee7caff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.457968] env[65788]: WARNING neutronclient.v2_0.client [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.458162] env[65788]: WARNING openstack [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.458772] env[65788]: WARNING openstack [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.491918] env[65788]: DEBUG oslo_vmware.api [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662318, 'name': PowerOnVM_Task, 'duration_secs': 0.524024} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.493545] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.495582] env[65788]: DEBUG nova.compute.manager [None req-1a068cef-51ed-45ee-92a0-a45d0734d585 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 790.496656] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7a65f5-dc7c-4854-a9e0-de649eae6c49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.505697] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 790.513955] env[65788]: DEBUG nova.network.neutron [-] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.542410] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 790.542759] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 790.542993] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 790.543293] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 790.543531] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 790.543756] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 790.544084] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.544336] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 790.544600] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 790.544842] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 790.545222] env[65788]: DEBUG nova.virt.hardware [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 790.546476] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da61ddf0-0638-48fd-a813-5213d7736c11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.559081] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0fe764-afa2-4e24-8ab6-db11679725cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.616739] env[65788]: WARNING openstack [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.617356] env[65788]: WARNING openstack [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.628304] env[65788]: INFO nova.compute.manager [-] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Took 1.37 seconds to deallocate network for instance. [ 790.662302] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.183s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 790.669673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 36.118s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 790.691457] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a6450f-99fa-2e63-38d9-6ad682b223bb, 'name': SearchDatastore_Task, 'duration_secs': 0.01103} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.691777] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.692063] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2c24bc30-d413-4714-81c2-b657cafe94bd/2c24bc30-d413-4714-81c2-b657cafe94bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.692364] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f249d2f-ae99-4c3f-a156-76da7ef7dea6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.696068] env[65788]: INFO nova.scheduler.client.report [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Deleted allocations for instance 72111766-217d-469f-ad92-ca0ce0e67090 [ 790.701917] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 790.701917] env[65788]: value = "task-4662319" [ 790.701917] env[65788]: _type = "Task" [ 790.701917] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.706581] env[65788]: WARNING neutronclient.v2_0.client [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 790.708223] env[65788]: WARNING openstack [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.708223] env[65788]: WARNING openstack [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.724420] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.815370] env[65788]: DEBUG nova.network.neutron [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 790.853873] env[65788]: INFO nova.compute.manager [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Took 41.65 seconds to build instance. [ 791.020831] env[65788]: INFO nova.compute.manager [-] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Took 1.32 seconds to deallocate network for instance. [ 791.065670] env[65788]: DEBUG nova.network.neutron [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Successfully updated port: 4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 791.138078] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.175639] env[65788]: INFO nova.compute.claims [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.209658] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1403ca6-17f0-4d09-85e8-215ab86fedc6 tempest-ServerAddressesNegativeTestJSON-497890566 tempest-ServerAddressesNegativeTestJSON-497890566-project-member] Lock "72111766-217d-469f-ad92-ca0ce0e67090" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.534s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.218335] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662319, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.319193] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.319632] env[65788]: DEBUG nova.objects.instance [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'migration_context' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 791.356443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0ce82a51-936b-4522-9863-5b3353e8690f tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "a813d102-1b77-4214-8eab-2cd66e002912" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.180s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.531187] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.569193] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "refresh_cache-47ff3955-5d9b-4d90-b8b1-276a0ce58a21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.569398] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired lock "refresh_cache-47ff3955-5d9b-4d90-b8b1-276a0ce58a21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 791.569588] env[65788]: DEBUG nova.network.neutron [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 791.672316] env[65788]: INFO nova.compute.manager [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Rebuilding instance [ 791.685598] env[65788]: INFO nova.compute.resource_tracker [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating resource usage from migration 894f791d-94eb-4f0f-b93a-e53dccce48e2 [ 791.719571] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662319, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519229} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.725759] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2c24bc30-d413-4714-81c2-b657cafe94bd/2c24bc30-d413-4714-81c2-b657cafe94bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.726940] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.728566] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94cf0473-7756-4870-903e-87e880e7d618 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.730740] env[65788]: DEBUG nova.compute.manager [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 791.731592] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bd3e7b-709b-43a7-b8ae-0a7c40d18700 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.748857] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 791.748857] env[65788]: value = "task-4662320" [ 791.748857] env[65788]: _type = "Task" [ 791.748857] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.759739] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662320, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.823067] env[65788]: DEBUG nova.objects.base [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 791.824426] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf05c61-11cc-4a3a-91d7-2a04d506369a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.856108] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4406f074-0fc6-4334-9dc0-33ac1afce01e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.864041] env[65788]: DEBUG oslo_vmware.api [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 791.864041] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522ac40b-3f5e-46a4-03ea-54a15dfd35bf" [ 791.864041] env[65788]: _type = "Task" [ 791.864041] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.875519] env[65788]: DEBUG oslo_vmware.api [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522ac40b-3f5e-46a4-03ea-54a15dfd35bf, 'name': SearchDatastore_Task, 'duration_secs': 0.008535} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.878424] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.964693] env[65788]: DEBUG nova.compute.manager [req-570c0328-b461-4a79-b6ec-6404f9794263 req-c93c6597-3222-44d7-b36f-33c1dec1f6cc service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Received event network-vif-plugged-4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 791.964796] env[65788]: DEBUG oslo_concurrency.lockutils [req-570c0328-b461-4a79-b6ec-6404f9794263 req-c93c6597-3222-44d7-b36f-33c1dec1f6cc service nova] Acquiring lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.965681] env[65788]: DEBUG oslo_concurrency.lockutils [req-570c0328-b461-4a79-b6ec-6404f9794263 req-c93c6597-3222-44d7-b36f-33c1dec1f6cc service nova] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.967125] env[65788]: DEBUG oslo_concurrency.lockutils [req-570c0328-b461-4a79-b6ec-6404f9794263 req-c93c6597-3222-44d7-b36f-33c1dec1f6cc service nova] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.967308] env[65788]: DEBUG nova.compute.manager [req-570c0328-b461-4a79-b6ec-6404f9794263 req-c93c6597-3222-44d7-b36f-33c1dec1f6cc service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] No waiting events found dispatching network-vif-plugged-4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 791.967481] env[65788]: WARNING nova.compute.manager [req-570c0328-b461-4a79-b6ec-6404f9794263 req-c93c6597-3222-44d7-b36f-33c1dec1f6cc service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Received unexpected event network-vif-plugged-4a920eec-6022-4912-93e4-46affc7b4d18 for instance with vm_state building and task_state spawning. [ 792.075116] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.075622] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.134135] env[65788]: DEBUG nova.network.neutron [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 792.202114] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.202586] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.265551] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662320, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073729} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.266316] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.267134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b7b571-1c33-4dfd-ae05-97d0227a2061 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.294625] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 2c24bc30-d413-4714-81c2-b657cafe94bd/2c24bc30-d413-4714-81c2-b657cafe94bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.300664] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-feec6a6a-3e06-498c-832f-0a2732d72d66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.331510] env[65788]: WARNING neutronclient.v2_0.client [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 792.332605] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.333264] env[65788]: WARNING openstack [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.358849] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 792.358849] env[65788]: value = "task-4662321" [ 792.358849] env[65788]: _type = "Task" [ 792.358849] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.372057] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662321, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.436971] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c4e234-82b7-4177-91da-624bc4813a78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.453159] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d0e358-1c34-45f4-b5d4-49407d8cc5b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.498282] env[65788]: DEBUG nova.network.neutron [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Updating instance_info_cache with network_info: [{"id": "4a920eec-6022-4912-93e4-46affc7b4d18", "address": "fa:16:3e:95:d7:8b", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a920eec-60", "ovs_interfaceid": "4a920eec-6022-4912-93e4-46affc7b4d18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.503203] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c1ba2e-f6b3-48a5-ab76-e708804f8832 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.512662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175455af-a7c1-449c-8b5d-01a2ef6b4bc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.533292] env[65788]: DEBUG nova.compute.provider_tree [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.757744] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 792.758128] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae7769f1-a10a-4753-89f0-c470f977b9dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.767193] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 792.767193] env[65788]: value = "task-4662322" [ 792.767193] env[65788]: _type = "Task" [ 792.767193] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.779143] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.874605] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662321, 'name': ReconfigVM_Task, 'duration_secs': 0.491257} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.874863] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 2c24bc30-d413-4714-81c2-b657cafe94bd/2c24bc30-d413-4714-81c2-b657cafe94bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 792.875919] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e89453c-d176-4cf9-bbf5-26d2712f0f95 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.886570] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 792.886570] env[65788]: value = "task-4662323" [ 792.886570] env[65788]: _type = "Task" [ 792.886570] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.902030] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662323, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.968043] env[65788]: DEBUG nova.compute.manager [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Received event network-vif-deleted-3a66ab74-5f68-4ab5-939e-de3cc3e69db2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 792.968195] env[65788]: DEBUG nova.compute.manager [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 792.968394] env[65788]: DEBUG nova.compute.manager [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing instance network info cache due to event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 792.968704] env[65788]: DEBUG oslo_concurrency.lockutils [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.969143] env[65788]: DEBUG oslo_concurrency.lockutils [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.969143] env[65788]: DEBUG nova.network.neutron [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 793.007358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Releasing lock "refresh_cache-47ff3955-5d9b-4d90-b8b1-276a0ce58a21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.007857] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Instance network_info: |[{"id": "4a920eec-6022-4912-93e4-46affc7b4d18", "address": "fa:16:3e:95:d7:8b", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a920eec-60", "ovs_interfaceid": "4a920eec-6022-4912-93e4-46affc7b4d18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 793.008517] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:d7:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92233552-2c0c-416e-9bf3-bfcca8eda2dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a920eec-6022-4912-93e4-46affc7b4d18', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.022451] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 793.023529] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 793.023908] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3aecf0da-8844-41ec-ae4f-db926df5c8dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.056158] env[65788]: DEBUG nova.scheduler.client.report [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 793.070168] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.070168] env[65788]: value = "task-4662324" [ 793.070168] env[65788]: _type = "Task" [ 793.070168] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.082035] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662324, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.278798] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662322, 'name': PowerOffVM_Task, 'duration_secs': 0.268715} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.279219] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 793.279347] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.280212] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c6936d-46fe-4f1f-8cdd-bf7387737dd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.288440] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 793.288793] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-98b71c43-7691-47a1-adc8-dde987b28806 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.320662] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 793.320846] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 793.320908] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Deleting the datastore file [datastore2] a813d102-1b77-4214-8eab-2cd66e002912 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.321242] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d962dea0-4fa3-4d56-ab1b-59cdce217bc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.329803] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 793.329803] env[65788]: value = "task-4662326" [ 793.329803] env[65788]: _type = "Task" [ 793.329803] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.342482] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.399167] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662323, 'name': Rename_Task, 'duration_secs': 0.245404} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.399548] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 793.399822] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8098d2f5-b27b-432e-91a3-1677bb1337fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.410199] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 793.410199] env[65788]: value = "task-4662327" [ 793.410199] env[65788]: _type = "Task" [ 793.410199] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.424666] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.472255] env[65788]: WARNING neutronclient.v2_0.client [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 793.473406] env[65788]: WARNING openstack [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.473983] env[65788]: WARNING openstack [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.567510] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.895s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.567510] env[65788]: INFO nova.compute.manager [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Migrating [ 793.574161] env[65788]: INFO nova.compute.manager [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Unrescuing [ 793.574498] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.574689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.574900] env[65788]: DEBUG nova.network.neutron [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 793.578749] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.973s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.579104] env[65788]: DEBUG nova.objects.instance [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lazy-loading 'resources' on Instance uuid 77dcec96-50e5-4753-b3cb-c6aec377bce2 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.599016] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662324, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.669855] env[65788]: WARNING openstack [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.670538] env[65788]: WARNING openstack [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.760415] env[65788]: WARNING neutronclient.v2_0.client [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 793.761497] env[65788]: WARNING openstack [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.761931] env[65788]: WARNING openstack [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.846931] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662326, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.333695} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.846931] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.846931] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 793.847220] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.901979] env[65788]: DEBUG nova.network.neutron [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updated VIF entry in instance network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 793.902217] env[65788]: DEBUG nova.network.neutron [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 793.922795] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662327, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.035509] env[65788]: DEBUG nova.compute.manager [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Received event network-changed-4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 794.036637] env[65788]: DEBUG nova.compute.manager [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Refreshing instance network info cache due to event network-changed-4a920eec-6022-4912-93e4-46affc7b4d18. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 794.036637] env[65788]: DEBUG oslo_concurrency.lockutils [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Acquiring lock "refresh_cache-47ff3955-5d9b-4d90-b8b1-276a0ce58a21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.036637] env[65788]: DEBUG oslo_concurrency.lockutils [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Acquired lock "refresh_cache-47ff3955-5d9b-4d90-b8b1-276a0ce58a21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.036637] env[65788]: DEBUG nova.network.neutron [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Refreshing network info cache for port 4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 794.086417] env[65788]: WARNING neutronclient.v2_0.client [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.091032] env[65788]: WARNING openstack [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.091032] env[65788]: WARNING openstack [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.099361] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.099917] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.099917] env[65788]: DEBUG nova.network.neutron [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 794.108520] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662324, 'name': CreateVM_Task, 'duration_secs': 0.758108} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.108520] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 794.108520] env[65788]: WARNING neutronclient.v2_0.client [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.108520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.108520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.109390] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 794.109390] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24788f53-16f4-480c-a098-a7f1afa3abb4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.115300] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 794.115300] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5241cdf9-9180-829d-89ed-5eee2d832dbc" [ 794.115300] env[65788]: _type = "Task" [ 794.115300] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.124906] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5241cdf9-9180-829d-89ed-5eee2d832dbc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.198301] env[65788]: WARNING openstack [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.198702] env[65788]: WARNING openstack [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.319335] env[65788]: WARNING neutronclient.v2_0.client [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.319967] env[65788]: WARNING openstack [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.320338] env[65788]: WARNING openstack [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.406284] env[65788]: DEBUG oslo_concurrency.lockutils [req-0f42ca86-d41e-49bd-a2d5-398cd3fd0ac3 req-9276c613-3437-4db7-8a74-6e43065171d7 service nova] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.424153] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662327, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.460554] env[65788]: DEBUG nova.network.neutron [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Updating instance_info_cache with network_info: [{"id": "f116c9d8-4c98-4af9-8747-0c585d99d738", "address": "fa:16:3e:e3:a3:1e", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf116c9d8-4c", "ovs_interfaceid": "f116c9d8-4c98-4af9-8747-0c585d99d738", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 794.539592] env[65788]: WARNING neutronclient.v2_0.client [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.540308] env[65788]: WARNING openstack [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.540666] env[65788]: WARNING openstack [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.605022] env[65788]: WARNING neutronclient.v2_0.client [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.605698] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.606173] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.629013] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5241cdf9-9180-829d-89ed-5eee2d832dbc, 'name': SearchDatastore_Task, 'duration_secs': 0.02044} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.632168] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.632426] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.632680] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.632848] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.633043] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.633777] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30eb99cf-b998-4038-9207-8cea67b57b89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.646795] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.646988] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 794.650066] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6660905f-c888-4cbc-8154-ca8edf7b715d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.653976] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 794.653976] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520e824c-c2a3-03fe-4432-0c36635c4368" [ 794.653976] env[65788]: _type = "Task" [ 794.653976] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.666802] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520e824c-c2a3-03fe-4432-0c36635c4368, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.687227] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2e0d69-a948-45f9-8b29-37bf8ab0a3f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.695716] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365f7ea8-dac1-47c6-820f-35626874360a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.734713] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1509a09-1a79-45e8-97c7-68f277fa02c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.744978] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26324c3b-20d7-41da-8266-f317a832a594 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.761434] env[65788]: DEBUG nova.compute.provider_tree [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.768099] env[65788]: WARNING openstack [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.768557] env[65788]: WARNING openstack [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.878485] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.878947] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.900012] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 794.900380] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 794.900569] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 794.900790] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 794.901022] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 794.901238] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 794.901515] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 794.901703] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 794.901914] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 794.902123] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 794.902315] env[65788]: DEBUG nova.virt.hardware [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 794.903352] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b602dcda-f288-4fe6-b160-d7282da901c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.916516] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbbc03a-083a-472b-8335-312f8cdaadb3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.928167] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662327, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.936997] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.943768] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 794.945365] env[65788]: WARNING neutronclient.v2_0.client [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.946013] env[65788]: WARNING openstack [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.946486] env[65788]: WARNING openstack [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.956867] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.957427] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ff80f12-8d5d-4008-9eff-3036f54c7ba8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.970975] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-9438ab56-1b4c-4778-a608-de319ab0ee43" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.971696] env[65788]: DEBUG nova.objects.instance [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'flavor' on Instance uuid 9438ab56-1b4c-4778-a608-de319ab0ee43 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 794.979997] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.979997] env[65788]: value = "task-4662328" [ 794.979997] env[65788]: _type = "Task" [ 794.979997] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.992272] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662328, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.099508] env[65788]: WARNING neutronclient.v2_0.client [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 795.103020] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.103020] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.129545] env[65788]: DEBUG nova.network.neutron [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Updated VIF entry in instance network info cache for port 4a920eec-6022-4912-93e4-46affc7b4d18. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 795.129926] env[65788]: DEBUG nova.network.neutron [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Updating instance_info_cache with network_info: [{"id": "4a920eec-6022-4912-93e4-46affc7b4d18", "address": "fa:16:3e:95:d7:8b", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.113", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a920eec-60", "ovs_interfaceid": "4a920eec-6022-4912-93e4-46affc7b4d18", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.166719] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520e824c-c2a3-03fe-4432-0c36635c4368, 'name': SearchDatastore_Task, 'duration_secs': 0.011204} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.167684] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa9d053c-1e1b-4a4a-8427-04f3f2747e0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.175574] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 795.175574] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee6582-e6ae-1833-d39d-cc1d4750ed65" [ 795.175574] env[65788]: _type = "Task" [ 795.175574] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.191140] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ee6582-e6ae-1833-d39d-cc1d4750ed65, 'name': SearchDatastore_Task, 'duration_secs': 0.01174} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.191682] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.191730] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 47ff3955-5d9b-4d90-b8b1-276a0ce58a21/47ff3955-5d9b-4d90-b8b1-276a0ce58a21.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 795.194047] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3ac8389-60ff-47ed-97cc-48608d22b688 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.202959] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 795.202959] env[65788]: value = "task-4662329" [ 795.202959] env[65788]: _type = "Task" [ 795.202959] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.214022] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662329, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.229293] env[65788]: DEBUG nova.network.neutron [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.265184] env[65788]: DEBUG nova.scheduler.client.report [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 795.422049] env[65788]: DEBUG oslo_vmware.api [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662327, 'name': PowerOnVM_Task, 'duration_secs': 1.60453} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.422450] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.422633] env[65788]: INFO nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Took 10.01 seconds to spawn the instance on the hypervisor. [ 795.422817] env[65788]: DEBUG nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 795.423672] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b614a3ed-feb5-46d4-8505-5b142fd4f6ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.480132] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278c3921-2ce4-4143-bcfc-3222b8032832 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.507619] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.511828] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a4366aa-9fe3-4a98-8a34-3a319c1705f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.513875] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662328, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.523629] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 795.523629] env[65788]: value = "task-4662330" [ 795.523629] env[65788]: _type = "Task" [ 795.523629] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.533731] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.633466] env[65788]: DEBUG oslo_concurrency.lockutils [req-7e6a52c0-2ac1-4627-a16c-008a9f851bbb req-42f87043-f3b5-4910-a06e-099efcce8011 service nova] Releasing lock "refresh_cache-47ff3955-5d9b-4d90-b8b1-276a0ce58a21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.717751] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662329, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.732830] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.770947] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.192s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.775954] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.688s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.778054] env[65788]: DEBUG nova.objects.instance [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lazy-loading 'resources' on Instance uuid 5b701040-025c-4246-ad54-f2cf478e998d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.819171] env[65788]: INFO nova.scheduler.client.report [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Deleted allocations for instance 77dcec96-50e5-4753-b3cb-c6aec377bce2 [ 795.955833] env[65788]: INFO nova.compute.manager [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Took 45.81 seconds to build instance. [ 795.999992] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662328, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.036611] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.213389] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662329, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555259} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.213680] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 47ff3955-5d9b-4d90-b8b1-276a0ce58a21/47ff3955-5d9b-4d90-b8b1-276a0ce58a21.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 796.213895] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.214184] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-60b19471-0b11-47d4-9d59-89ad5f153373 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.223378] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 796.223378] env[65788]: value = "task-4662331" [ 796.223378] env[65788]: _type = "Task" [ 796.223378] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.253922] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.331170] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd0c98f5-ca8b-4103-8b8f-378e6bed9c5b tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "77dcec96-50e5-4753-b3cb-c6aec377bce2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.670s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.457132] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fd73021-ac17-4c4d-bbf3-66b5333de0ce tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.866s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.497443] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662328, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.536157] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.742277] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094047} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.742629] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.743535] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832ff20f-b21e-4bd6-a4ca-8c8275208c92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.780939] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 47ff3955-5d9b-4d90-b8b1-276a0ce58a21/47ff3955-5d9b-4d90-b8b1-276a0ce58a21.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.787093] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4f71b8d-8e66-4dd7-a93a-afed5ebf3bf7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.803448] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98f756d-3557-43fc-9070-d5e46ce13864 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.826217] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 796.834856] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 796.834856] env[65788]: value = "task-4662332" [ 796.834856] env[65788]: _type = "Task" [ 796.834856] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.846503] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.996961] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c144d095-be4e-4465-ae95-f11c8590ca76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.009516] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662328, 'name': CreateVM_Task, 'duration_secs': 1.568358} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.011733] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.012347] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.012503] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.012817] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 797.013853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a77839e-6956-43d6-9182-b0ba6b427d83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.017327] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7f54813-a69b-49e1-9869-d903520654ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.024715] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 797.024715] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5200f23e-66d6-1c32-900d-6b68676b64bb" [ 797.024715] env[65788]: _type = "Task" [ 797.024715] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.060908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98138c81-b9cf-4f2c-a11f-8bfc8920b7b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.076879] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c260c4-845b-4574-9196-7ee66fa445d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.081102] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662330, 'name': PowerOffVM_Task, 'duration_secs': 1.131733} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.081703] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5200f23e-66d6-1c32-900d-6b68676b64bb, 'name': SearchDatastore_Task, 'duration_secs': 0.020509} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.081963] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.087349] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Reconfiguring VM instance instance-0000002c to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 797.087813] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.087909] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.088165] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.088312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.088492] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.089253] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50aafe2b-502e-4c9e-9161-e75b1d199c10 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.102448] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33c9808c-d372-46d1-917a-79aff7bb6e40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.112759] env[65788]: DEBUG nova.compute.provider_tree [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.121484] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 797.121484] env[65788]: value = "task-4662333" [ 797.121484] env[65788]: _type = "Task" [ 797.121484] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.125951] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.126147] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.129090] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ac93b95-fb6c-44d4-a045-7fad6d51b7fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.136874] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662333, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.138411] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 797.138411] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525eac82-4f16-48be-8967-d69dd3dc2379" [ 797.138411] env[65788]: _type = "Task" [ 797.138411] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.150836] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525eac82-4f16-48be-8967-d69dd3dc2379, 'name': SearchDatastore_Task, 'duration_secs': 0.010385} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.152135] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58ca19bd-3d05-4b52-9cfe-6179df8d2211 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.158663] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 797.158663] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5265c03f-0636-d6ad-12ab-bfde35e38e44" [ 797.158663] env[65788]: _type = "Task" [ 797.158663] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.169271] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5265c03f-0636-d6ad-12ab-bfde35e38e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.337523] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.338013] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ed21bd0-b5d0-465c-9468-5fc509142e0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.351372] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662332, 'name': ReconfigVM_Task, 'duration_secs': 0.403639} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.353186] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 47ff3955-5d9b-4d90-b8b1-276a0ce58a21/47ff3955-5d9b-4d90-b8b1-276a0ce58a21.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.354055] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 797.354055] env[65788]: value = "task-4662334" [ 797.354055] env[65788]: _type = "Task" [ 797.354055] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.354318] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-404e3af1-c5b2-40fe-98a6-71cd9913be14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.371197] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.371603] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 797.371603] env[65788]: value = "task-4662335" [ 797.371603] env[65788]: _type = "Task" [ 797.371603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.619022] env[65788]: DEBUG nova.scheduler.client.report [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 797.643081] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.643366] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.645096] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662333, 'name': ReconfigVM_Task, 'duration_secs': 0.445644} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.647024] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Reconfigured VM instance instance-0000002c to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 797.647220] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.647756] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e1f92ff-d368-46da-899a-91e375d040f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.661044] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 797.661044] env[65788]: value = "task-4662336" [ 797.661044] env[65788]: _type = "Task" [ 797.661044] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.678136] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5265c03f-0636-d6ad-12ab-bfde35e38e44, 'name': SearchDatastore_Task, 'duration_secs': 0.014732} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.678136] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.679611] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.679611] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 797.680393] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4e882dc-f1d3-4b94-bf6e-abdbd4d1bac5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.692266] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 797.692266] env[65788]: value = "task-4662337" [ 797.692266] env[65788]: _type = "Task" [ 797.692266] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.703958] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.869894] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662334, 'name': PowerOffVM_Task, 'duration_secs': 0.240492} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.870314] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.870545] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 797.890197] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662335, 'name': Rename_Task, 'duration_secs': 0.18237} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.890669] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.890961] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de29bdf0-9348-49c5-84cf-820c777c850d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.900205] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 797.900205] env[65788]: value = "task-4662338" [ 797.900205] env[65788]: _type = "Task" [ 797.900205] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.912380] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662338, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.127302] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.353s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.130727] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.212s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.131023] env[65788]: DEBUG nova.objects.instance [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lazy-loading 'resources' on Instance uuid 6aa4b17b-816c-4d84-8f74-a81185f3af65 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.146888] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 798.156753] env[65788]: INFO nova.scheduler.client.report [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Deleted allocations for instance 5b701040-025c-4246-ad54-f2cf478e998d [ 798.175201] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662336, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.205766] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.384039] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 798.384295] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 798.384991] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 798.384991] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 798.384991] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 798.384991] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 798.385258] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.385258] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 798.385505] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 798.385671] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 798.385836] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 798.391073] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36e22342-caf3-4a02-91cf-6d484aa5de45 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.411528] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662338, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.413705] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 798.413705] env[65788]: value = "task-4662339" [ 798.413705] env[65788]: _type = "Task" [ 798.413705] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.423447] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662339, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.676569] env[65788]: DEBUG oslo_vmware.api [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662336, 'name': PowerOnVM_Task, 'duration_secs': 0.666704} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.679542] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 798.679806] env[65788]: DEBUG nova.compute.manager [None req-a2ea8316-1eed-4308-aec2-6212199d478c tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 798.680413] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9dc7c376-1182-4c1d-b0f4-16c0e060fb05 tempest-VolumesAssistedSnapshotsTest-1577585050 tempest-VolumesAssistedSnapshotsTest-1577585050-project-member] Lock "5b701040-025c-4246-ad54-f2cf478e998d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 45.074s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.684397] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41666b66-04be-4ba2-a6b1-a52965183c22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.688033] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.709634] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662337, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.923307] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662338, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.934921] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662339, 'name': ReconfigVM_Task, 'duration_secs': 0.246563} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.935319] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 799.202847] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6b2a5b-a40e-46fc-aff2-5bf4a49ebe15 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.216338] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26549f3c-9e59-4fca-8d4d-51f54e8265ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.219990] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662337, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.311772} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.222817] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.222817] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.222817] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2161da4-3e38-40a9-8413-60d21e59395c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.257081] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bac04b-2718-4455-a234-2056c5d39d02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.260596] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 799.260596] env[65788]: value = "task-4662340" [ 799.260596] env[65788]: _type = "Task" [ 799.260596] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.273094] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52776274-374a-49e5-bf86-be916ce37407 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.281655] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662340, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.293910] env[65788]: DEBUG nova.compute.provider_tree [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.414737] env[65788]: DEBUG oslo_vmware.api [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662338, 'name': PowerOnVM_Task, 'duration_secs': 1.414696} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.415068] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.415308] env[65788]: INFO nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Took 8.91 seconds to spawn the instance on the hypervisor. [ 799.415495] env[65788]: DEBUG nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 799.416291] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a7610f-319e-4fc8-bdc5-caed23be2d52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.442614] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:47:39Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='edeb65e9-37d7-4c27-800f-7029f9e8ded0',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1773652449',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 799.442897] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 799.443125] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 799.443376] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 799.443557] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 799.443983] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 799.444330] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.444540] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 799.445021] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 799.445387] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 799.445616] env[65788]: DEBUG nova.virt.hardware [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 799.452429] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfiguring VM instance instance-00000028 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 799.452696] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-931656cf-da44-4e78-91ad-bac535327c50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.474070] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 799.474070] env[65788]: value = "task-4662341" [ 799.474070] env[65788]: _type = "Task" [ 799.474070] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.484081] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662341, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.771042] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078634} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.771392] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.772180] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086c2728-5f37-4a1d-99d1-efc97f77d003 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.793019] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.793339] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-930e7972-db4d-4f34-9eb6-f32a1c1edd25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.809637] env[65788]: DEBUG nova.scheduler.client.report [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 799.820059] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 799.820059] env[65788]: value = "task-4662342" [ 799.820059] env[65788]: _type = "Task" [ 799.820059] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.830759] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662342, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.943480] env[65788]: INFO nova.compute.manager [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Took 48.89 seconds to build instance. [ 799.990373] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662341, 'name': ReconfigVM_Task, 'duration_secs': 0.473582} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.991249] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfigured VM instance instance-00000028 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 799.992315] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d644ee5-8e8d-4fd0-aa90-b6793b178258 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.017850] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.018206] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e8847e1-e716-4d6d-9463-46e33178fce8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.038807] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 800.038807] env[65788]: value = "task-4662343" [ 800.038807] env[65788]: _type = "Task" [ 800.038807] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.048505] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662343, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.315732] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.185s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.318523] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 40.585s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.318756] env[65788]: DEBUG nova.objects.instance [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lazy-loading 'resources' on Instance uuid 6be541bd-06fd-4959-9987-d250011b3f1d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.336230] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662342, 'name': ReconfigVM_Task, 'duration_secs': 0.471094} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.337439] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Reconfigured VM instance instance-00000032 to attach disk [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.338491] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ebaa0cd0-9a8b-4086-8e88-d4431be3c1f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.348984] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 800.348984] env[65788]: value = "task-4662344" [ 800.348984] env[65788]: _type = "Task" [ 800.348984] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.350331] env[65788]: INFO nova.scheduler.client.report [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Deleted allocations for instance 6aa4b17b-816c-4d84-8f74-a81185f3af65 [ 800.364683] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662344, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.445210] env[65788]: DEBUG oslo_concurrency.lockutils [None req-54969e91-4142-4484-b4bf-42d615cbd979 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.940s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.551128] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662343, 'name': ReconfigVM_Task, 'duration_secs': 0.401541} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.551128] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfigured VM instance instance-00000028 to attach disk [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.551128] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 800.864192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fc31d7-6c82-47e1-99b6-7d911353649c tempest-ServerMetadataTestJSON-2105857083 tempest-ServerMetadataTestJSON-2105857083-project-member] Lock "6aa4b17b-816c-4d84-8f74-a81185f3af65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 46.217s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.875126] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662344, 'name': Rename_Task, 'duration_secs': 0.167889} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.876122] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.880132] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a83aabf6-7f77-475a-b996-6ce7cbff1550 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.892562] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 800.892562] env[65788]: value = "task-4662345" [ 800.892562] env[65788]: _type = "Task" [ 800.892562] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.907409] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.062337] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b3eda0-d448-4f32-8510-927b776b411e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.088270] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0f69d6-e24c-4506-b001-49d5a0ec6612 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.108171] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 801.405479] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662345, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.407245] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbdfa37-3df3-407a-9477-c85cf4a5edd3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.414938] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad457fd-b319-474d-9981-48ef80302c70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.447150] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e00fef2-c52d-44bf-80d6-4e0693112910 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.456112] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa58b9da-b7ce-43de-80ff-a572da6f2aac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.472514] env[65788]: DEBUG nova.compute.provider_tree [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.514622] env[65788]: DEBUG nova.compute.manager [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 801.516258] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8962ab4c-473d-4dc1-b151-481491367d3a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.618107] env[65788]: WARNING neutronclient.v2_0.client [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.906637] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662345, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.976162] env[65788]: DEBUG nova.scheduler.client.report [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.030238] env[65788]: INFO nova.compute.manager [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] instance snapshotting [ 802.035584] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f6a6fa-dab0-4b38-ad3d-3826e8cca1a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.060980] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9e159a-7028-4743-9e0b-3405be70b620 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.246777] env[65788]: DEBUG nova.network.neutron [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Port 04994181-be30-4c51-b4a1-ce198600a5c3 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 802.408552] env[65788]: DEBUG oslo_vmware.api [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662345, 'name': PowerOnVM_Task, 'duration_secs': 1.328566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.409107] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 802.409107] env[65788]: DEBUG nova.compute.manager [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 802.409874] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0212d08d-9fc9-4a71-9993-a06efdff1b95 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.482861] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.164s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.486375] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.684s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.488722] env[65788]: INFO nova.compute.claims [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.514232] env[65788]: INFO nova.scheduler.client.report [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Deleted allocations for instance 6be541bd-06fd-4959-9987-d250011b3f1d [ 802.579023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 802.579023] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c87ce873-f430-4871-8d46-b9df918efae1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.591699] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 802.591699] env[65788]: value = "task-4662346" [ 802.591699] env[65788]: _type = "Task" [ 802.591699] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.604261] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662346, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.931539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.025526] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a00bc47-3a42-4254-8192-202c38fb8420 tempest-InstanceActionsTestJSON-1234363326 tempest-InstanceActionsTestJSON-1234363326-project-member] Lock "6be541bd-06fd-4959-9987-d250011b3f1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 47.295s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.100767] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662346, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.271329] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.271769] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.271909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.609682] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662346, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.105975] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f21f9f-446f-4ec4-9698-eaa6f22c551a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.112281] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662346, 'name': CreateSnapshot_Task, 'duration_secs': 1.177079} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.113238] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 804.113859] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c55137-2a83-40b1-9e59-d41b7dc88bc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.123800] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9832eb1-1aed-4486-bccc-caa656ceea0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.163262] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4f2992-bf93-4c9d-8f69-5103d8c87f31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.171938] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b2b314-7a1b-4f99-beaf-529fb89614e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.190441] env[65788]: DEBUG nova.compute.provider_tree [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.251936] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.253213] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.277831] env[65788]: WARNING neutronclient.v2_0.client [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 804.583757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.583862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.585021] env[65788]: DEBUG nova.network.neutron [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 804.640963] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 804.641334] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9e9a6e00-5689-444b-88e6-8f10e6114228 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.651939] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 804.651939] env[65788]: value = "task-4662347" [ 804.651939] env[65788]: _type = "Task" [ 804.651939] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.663478] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662347, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.692145] env[65788]: DEBUG nova.scheduler.client.report [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 804.756137] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 805.088976] env[65788]: WARNING neutronclient.v2_0.client [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.088976] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.088976] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.166201] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662347, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.199582] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.713s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.200204] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 805.207056] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 43.183s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.207285] env[65788]: DEBUG nova.objects.instance [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 805.295594] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.328062] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.328398] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.436281] env[65788]: WARNING neutronclient.v2_0.client [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.437673] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.438344] env[65788]: WARNING openstack [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.579397] env[65788]: DEBUG nova.network.neutron [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.672646] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662347, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.709362] env[65788]: DEBUG nova.compute.utils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 805.710874] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 805.712114] env[65788]: DEBUG nova.network.neutron [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 805.715021] env[65788]: WARNING neutronclient.v2_0.client [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.715021] env[65788]: WARNING neutronclient.v2_0.client [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 805.715021] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 805.715021] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 805.807439] env[65788]: DEBUG nova.policy [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32b4adcc8fa42fba0dbffd4582fcf69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36859c1eb994614b2a77400f811cf50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 806.086066] env[65788]: DEBUG oslo_concurrency.lockutils [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.170956] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662347, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.230040] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 806.234183] env[65788]: DEBUG oslo_concurrency.lockutils [None req-453fa726-8457-41d6-a140-b7a3e6b2d5bb tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.027s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.235140] env[65788]: INFO nova.compute.manager [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Rebuilding instance [ 806.238103] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.462s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.238978] env[65788]: INFO nova.compute.claims [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.247311] env[65788]: DEBUG nova.network.neutron [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Successfully created port: 29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 806.326920] env[65788]: DEBUG nova.compute.manager [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 806.327860] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9ae293-fbff-4f50-b950-989cad36881c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.617233] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43e0f11-0b39-437d-a47d-0e2982a25536 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.643209] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5da596a-e861-451d-bf3d-474df986b2a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.653190] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 806.667119] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662347, 'name': CloneVM_Task, 'duration_secs': 1.788006} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.667463] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Created linked-clone VM from snapshot [ 806.668790] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22453747-03b0-46c3-b99a-0934ca0291ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.678207] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Uploading image 2ab5928d-b287-4fb9-8e31-7911bdbeafc7 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 806.720278] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 806.720278] env[65788]: value = "vm-910271" [ 806.720278] env[65788]: _type = "VirtualMachine" [ 806.720278] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 806.720600] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b8182518-b822-4a43-9d24-c963b4c838e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.729128] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lease: (returnval){ [ 806.729128] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528a08d2-ed03-3a8d-f61e-e3c3f5f8aa57" [ 806.729128] env[65788]: _type = "HttpNfcLease" [ 806.729128] env[65788]: } obtained for exporting VM: (result){ [ 806.729128] env[65788]: value = "vm-910271" [ 806.729128] env[65788]: _type = "VirtualMachine" [ 806.729128] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 806.729459] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the lease: (returnval){ [ 806.729459] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528a08d2-ed03-3a8d-f61e-e3c3f5f8aa57" [ 806.729459] env[65788]: _type = "HttpNfcLease" [ 806.729459] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 806.741236] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 806.741236] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528a08d2-ed03-3a8d-f61e-e3c3f5f8aa57" [ 806.741236] env[65788]: _type = "HttpNfcLease" [ 806.741236] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 807.163834] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 807.164190] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44a78896-da84-43ce-b68e-d7399ff16967 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.173184] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 807.173184] env[65788]: value = "task-4662349" [ 807.173184] env[65788]: _type = "Task" [ 807.173184] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.183723] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662349, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.239259] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 807.239259] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528a08d2-ed03-3a8d-f61e-e3c3f5f8aa57" [ 807.239259] env[65788]: _type = "HttpNfcLease" [ 807.239259] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 807.239603] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 807.239603] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528a08d2-ed03-3a8d-f61e-e3c3f5f8aa57" [ 807.239603] env[65788]: _type = "HttpNfcLease" [ 807.239603] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 807.240565] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d2b53a-5aaf-4f2a-87ab-9ede46028f59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.244193] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 807.252910] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f56b82-b60b-7bcf-e003-6613bc07df66/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 807.253191] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f56b82-b60b-7bcf-e003-6613bc07df66/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 807.326654] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 807.326926] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 807.327191] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 807.327320] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 807.327472] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 807.327628] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 807.327847] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.328031] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 807.328213] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 807.328392] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 807.328577] env[65788]: DEBUG nova.virt.hardware [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 807.329597] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84885bb-2db2-4b20-8edc-f132fb08adc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.341801] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932bc533-6726-4ac5-8a7f-c59313ca8e15 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.350241] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.350506] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c2b9f20-d654-42fa-a29d-d78e15f124a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.378110] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 807.378110] env[65788]: value = "task-4662350" [ 807.378110] env[65788]: _type = "Task" [ 807.378110] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.384104] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-359c35d6-908b-45b4-9c1e-9868d9311b22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.393350] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.690161] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662349, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.896228] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662350, 'name': PowerOffVM_Task, 'duration_secs': 0.335329} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.903942] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 807.903942] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.907541] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246098a5-c74f-4c29-9169-6652be1f17bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.921669] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 807.925929] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b27ef10-2db8-495f-9fd3-3eaeda5cd211 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.955508] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 807.955939] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 807.956187] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Deleting the datastore file [datastore2] a813d102-1b77-4214-8eab-2cd66e002912 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.958393] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a5641a8-76eb-4515-b05f-a92c9849861c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.961876] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc13115-245a-4f15-88ae-b2151f0b6863 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.972471] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 807.972471] env[65788]: value = "task-4662352" [ 807.972471] env[65788]: _type = "Task" [ 807.972471] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.980176] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1608fcb-4ff6-4ad2-8033-753c94124a8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.022954] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25badc9a-b6f3-4134-807e-83e56131ea37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.026935] env[65788]: DEBUG nova.compute.manager [req-b72d0873-023f-4509-a5ab-b293d41fcca4 req-6a1262e7-14fc-48d0-92c1-cb072cdd81e2 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Received event network-vif-plugged-29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 808.027267] env[65788]: DEBUG oslo_concurrency.lockutils [req-b72d0873-023f-4509-a5ab-b293d41fcca4 req-6a1262e7-14fc-48d0-92c1-cb072cdd81e2 service nova] Acquiring lock "caceb296-5da2-4b4d-b955-5d7a238ff939-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 808.027539] env[65788]: DEBUG oslo_concurrency.lockutils [req-b72d0873-023f-4509-a5ab-b293d41fcca4 req-6a1262e7-14fc-48d0-92c1-cb072cdd81e2 service nova] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 808.027730] env[65788]: DEBUG oslo_concurrency.lockutils [req-b72d0873-023f-4509-a5ab-b293d41fcca4 req-6a1262e7-14fc-48d0-92c1-cb072cdd81e2 service nova] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.027920] env[65788]: DEBUG nova.compute.manager [req-b72d0873-023f-4509-a5ab-b293d41fcca4 req-6a1262e7-14fc-48d0-92c1-cb072cdd81e2 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] No waiting events found dispatching network-vif-plugged-29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 808.028085] env[65788]: WARNING nova.compute.manager [req-b72d0873-023f-4509-a5ab-b293d41fcca4 req-6a1262e7-14fc-48d0-92c1-cb072cdd81e2 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Received unexpected event network-vif-plugged-29e1dcd1-5465-47b7-929e-de731202dbd0 for instance with vm_state building and task_state spawning. [ 808.036868] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa61131a-06b4-4c29-982c-e3507dd8557e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.057032] env[65788]: DEBUG nova.compute.provider_tree [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.082773] env[65788]: DEBUG nova.network.neutron [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Successfully updated port: 29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 808.188582] env[65788]: DEBUG oslo_vmware.api [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662349, 'name': PowerOnVM_Task, 'duration_secs': 0.589366} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.189653] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.189653] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-db2e7741-8c9f-4214-8f7f-96c1cfe37cac tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance 'f4a490a2-f2b1-4eac-8c1a-a18758583c70' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 808.485309] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112478} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.485738] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.486104] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.486396] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.560695] env[65788]: DEBUG nova.scheduler.client.report [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 808.587266] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-caceb296-5da2-4b4d-b955-5d7a238ff939" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.587530] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-caceb296-5da2-4b4d-b955-5d7a238ff939" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.588293] env[65788]: DEBUG nova.network.neutron [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 809.066295] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.829s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.068351] env[65788]: DEBUG nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 809.073152] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.606s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.073152] env[65788]: DEBUG nova.objects.instance [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lazy-loading 'resources' on Instance uuid cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.090830] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.091501] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.150215] env[65788]: DEBUG nova.network.neutron [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 809.173254] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.173254] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.325643] env[65788]: WARNING neutronclient.v2_0.client [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 809.328502] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.328974] env[65788]: WARNING openstack [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.453173] env[65788]: DEBUG nova.network.neutron [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Updating instance_info_cache with network_info: [{"id": "29e1dcd1-5465-47b7-929e-de731202dbd0", "address": "fa:16:3e:c5:e7:6b", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e1dcd1-54", "ovs_interfaceid": "29e1dcd1-5465-47b7-929e-de731202dbd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.521078] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 809.521355] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 809.521729] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 809.521971] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 809.522603] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 809.522860] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 809.523169] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.523339] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 809.524044] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 809.524268] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 809.524450] env[65788]: DEBUG nova.virt.hardware [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 809.525421] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c71ee5-7386-4ccf-93a8-2fd8d2918540 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.535174] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129b0b7c-db17-450e-8dee-1580567e64e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.551862] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.558654] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 809.558954] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.559216] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-052ab31c-e396-4e7a-9191-59b5af5f3fb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.575516] env[65788]: DEBUG nova.compute.utils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 809.580703] env[65788]: DEBUG nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 809.587290] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.587290] env[65788]: value = "task-4662353" [ 809.587290] env[65788]: _type = "Task" [ 809.587290] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.596772] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662353, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.964599] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-caceb296-5da2-4b4d-b955-5d7a238ff939" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.964922] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Instance network_info: |[{"id": "29e1dcd1-5465-47b7-929e-de731202dbd0", "address": "fa:16:3e:c5:e7:6b", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e1dcd1-54", "ovs_interfaceid": "29e1dcd1-5465-47b7-929e-de731202dbd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 809.967030] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e7:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '29e1dcd1-5465-47b7-929e-de731202dbd0', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.976116] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating folder: Project (e36859c1eb994614b2a77400f811cf50). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.977754] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d766ac5-cd8e-449a-abca-cb759311c208 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.003292] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created folder: Project (e36859c1eb994614b2a77400f811cf50) in parent group-v910111. [ 810.003292] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating folder: Instances. Parent ref: group-v910273. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.006148] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94b63bff-48ef-4c36-a8b5-73273d7338b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.020987] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created folder: Instances in parent group-v910273. [ 810.021121] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 810.021369] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 810.021672] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd0ac090-2d0c-4941-bd27-e70c01632490 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.050965] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.050965] env[65788]: value = "task-4662356" [ 810.050965] env[65788]: _type = "Task" [ 810.050965] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.063910] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662356, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.083148] env[65788]: DEBUG nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 810.106875] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662353, 'name': CreateVM_Task, 'duration_secs': 0.49525} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.108579] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.108776] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.109616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.109616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 810.110389] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d44a84bc-a4f1-4ba3-9f2b-1872576bfa67 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.121094] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 810.121094] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5253ab66-e586-a3cd-9373-23de25cc492b" [ 810.121094] env[65788]: _type = "Task" [ 810.121094] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.133221] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5253ab66-e586-a3cd-9373-23de25cc492b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.175782] env[65788]: DEBUG nova.compute.manager [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Received event network-changed-29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 810.176376] env[65788]: DEBUG nova.compute.manager [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Refreshing instance network info cache due to event network-changed-29e1dcd1-5465-47b7-929e-de731202dbd0. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 810.176468] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Acquiring lock "refresh_cache-caceb296-5da2-4b4d-b955-5d7a238ff939" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.176935] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Acquired lock "refresh_cache-caceb296-5da2-4b4d-b955-5d7a238ff939" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.176935] env[65788]: DEBUG nova.network.neutron [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Refreshing network info cache for port 29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 810.244087] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cd3474-8521-45ce-ab52-0dfd238415cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.253256] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171f9364-54b0-45b2-9eb7-2b333946b320 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.287263] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17850666-ba1a-42e0-ba82-80a64041e6ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.296486] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22edb362-648a-4ef5-a4b5-7a8407bb238c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.314828] env[65788]: DEBUG nova.compute.provider_tree [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.493459] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.532073] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.532073] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.563591] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662356, 'name': CreateVM_Task, 'duration_secs': 0.426191} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.563591] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.564066] env[65788]: WARNING neutronclient.v2_0.client [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.564546] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.573292] env[65788]: DEBUG nova.network.neutron [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Port 04994181-be30-4c51-b4a1-ce198600a5c3 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 810.574013] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.574013] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.574013] env[65788]: DEBUG nova.network.neutron [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 810.635685] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5253ab66-e586-a3cd-9373-23de25cc492b, 'name': SearchDatastore_Task, 'duration_secs': 0.015996} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.636132] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.638283] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.638283] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.638283] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.638283] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 810.638656] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 810.638656] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 810.638656] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f6d81d1e-99db-4f66-b06a-7eb37f9e2a91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.641108] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef8291b1-3905-46cf-bc90-5585a3a27b23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.649470] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 810.649470] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204bab4-2685-aa05-2e29-daa9a8c09740" [ 810.649470] env[65788]: _type = "Task" [ 810.649470] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.657200] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 810.657200] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 810.658115] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a5e1c91-b10f-42cd-8537-98b27e8422c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.667601] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204bab4-2685-aa05-2e29-daa9a8c09740, 'name': SearchDatastore_Task, 'duration_secs': 0.010698} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.668359] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.668752] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 810.669061] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.672510] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 810.672510] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff7338-5e36-2930-1efd-910af5c2c59e" [ 810.672510] env[65788]: _type = "Task" [ 810.672510] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.681971] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff7338-5e36-2930-1efd-910af5c2c59e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.683702] env[65788]: WARNING neutronclient.v2_0.client [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.684270] env[65788]: WARNING openstack [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.684619] env[65788]: WARNING openstack [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.818919] env[65788]: DEBUG nova.scheduler.client.report [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 810.835300] env[65788]: WARNING openstack [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.835747] env[65788]: WARNING openstack [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 810.901172] env[65788]: WARNING neutronclient.v2_0.client [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 810.901895] env[65788]: WARNING openstack [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 810.902258] env[65788]: WARNING openstack [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.026107] env[65788]: DEBUG nova.network.neutron [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Updated VIF entry in instance network info cache for port 29e1dcd1-5465-47b7-929e-de731202dbd0. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 811.026507] env[65788]: DEBUG nova.network.neutron [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Updating instance_info_cache with network_info: [{"id": "29e1dcd1-5465-47b7-929e-de731202dbd0", "address": "fa:16:3e:c5:e7:6b", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap29e1dcd1-54", "ovs_interfaceid": "29e1dcd1-5465-47b7-929e-de731202dbd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 811.076313] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 811.077035] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.077462] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.100343] env[65788]: DEBUG nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 811.128186] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 811.128444] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 811.128619] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 811.128789] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 811.130283] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 811.130841] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 811.130841] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 811.131073] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 811.131123] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 811.131293] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 811.131540] env[65788]: DEBUG nova.virt.hardware [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 811.132517] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb76810a-9978-45aa-84f5-4c3e8e2361aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.143942] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8fb6dc-521a-4e2e-8482-9b98f9fe364d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.161149] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.167873] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Creating folder: Project (0c9188d7fe304d97ac0882b39b9ae449). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.172279] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f587902-e31b-4fad-98ca-efb0d9c2b9bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.185669] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff7338-5e36-2930-1efd-910af5c2c59e, 'name': SearchDatastore_Task, 'duration_secs': 0.011011} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.186539] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d88e063-5bc4-4f95-b238-547dd897eee1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.190758] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Created folder: Project (0c9188d7fe304d97ac0882b39b9ae449) in parent group-v910111. [ 811.190758] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Creating folder: Instances. Parent ref: group-v910276. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.191173] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff32cf84-d225-4063-b904-95f66ee9ccc9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.194803] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 811.194803] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae4014-2b84-51cc-9ef9-171abfc4311b" [ 811.194803] env[65788]: _type = "Task" [ 811.194803] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.199474] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.199713] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.208173] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Created folder: Instances in parent group-v910276. [ 811.208443] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 811.208732] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.208931] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad333e57-683b-4774-8d64-fa2865f7f59a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.226942] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae4014-2b84-51cc-9ef9-171abfc4311b, 'name': SearchDatastore_Task, 'duration_secs': 0.011007} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.227676] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.227942] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.228282] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.228477] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.228704] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76a6b2d6-24b2-420f-96de-3a5321514928 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.231078] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99251a73-8665-4414-a70e-76d5546176a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.234741] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.234741] env[65788]: value = "task-4662359" [ 811.234741] env[65788]: _type = "Task" [ 811.234741] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.244087] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 811.244087] env[65788]: value = "task-4662360" [ 811.244087] env[65788]: _type = "Task" [ 811.244087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.246061] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.246220] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.250078] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca0c6816-2adc-4e41-b0ea-6ab0da011bd5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.256190] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662359, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.262394] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662360, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.263959] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 811.263959] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3ecea-57fb-aeb6-1617-d25bce1dd5ba" [ 811.263959] env[65788]: _type = "Task" [ 811.263959] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.275154] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3ecea-57fb-aeb6-1617-d25bce1dd5ba, 'name': SearchDatastore_Task, 'duration_secs': 0.01133} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.276244] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2fa779f-16a6-4cbc-bc65-7416a015fc57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.283376] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 811.283376] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a91aa8-37ff-14c5-6fdd-74d47b108d7f" [ 811.283376] env[65788]: _type = "Task" [ 811.283376] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.293295] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a91aa8-37ff-14c5-6fdd-74d47b108d7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.324137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.252s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 811.329155] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.945s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 811.329155] env[65788]: INFO nova.compute.claims [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.339032] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 811.339032] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.339032] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.354619] env[65788]: INFO nova.scheduler.client.report [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Deleted allocations for instance cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e [ 811.528658] env[65788]: DEBUG nova.network.neutron [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 811.530576] env[65788]: DEBUG oslo_concurrency.lockutils [req-d4b30b5a-7953-4f08-9124-bcd9cda1113e req-cac026fb-9632-45b2-b2f5-871018764019 service nova] Releasing lock "refresh_cache-caceb296-5da2-4b4d-b955-5d7a238ff939" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.746627] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662359, 'name': CreateVM_Task, 'duration_secs': 0.391472} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.753022] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.753022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.753022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.753022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 811.753022] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c9ddc09-8ef0-475c-b17b-523726fdb1f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.757508] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662360, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.762598] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 811.762598] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52080c9e-f15c-d5ab-afd6-2f0c94e45cef" [ 811.762598] env[65788]: _type = "Task" [ 811.762598] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.771921] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52080c9e-f15c-d5ab-afd6-2f0c94e45cef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.794265] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a91aa8-37ff-14c5-6fdd-74d47b108d7f, 'name': SearchDatastore_Task, 'duration_secs': 0.011219} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.794626] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.794938] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] caceb296-5da2-4b4d-b955-5d7a238ff939/caceb296-5da2-4b4d-b955-5d7a238ff939.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 811.795226] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6acd0f2-790c-4889-a78c-89f6f3f78f4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.802919] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 811.802919] env[65788]: value = "task-4662361" [ 811.802919] env[65788]: _type = "Task" [ 811.802919] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.817052] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662361, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.867491] env[65788]: DEBUG oslo_concurrency.lockutils [None req-209048be-e05b-4493-bc52-b3533537bd2b tempest-ServerShowV254Test-1430518618 tempest-ServerShowV254Test-1430518618-project-member] Lock "cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.721s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.036736] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.259536] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662360, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519928} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.259900] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.260141] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.260419] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f96febe6-b633-4e39-8b1b-db3332d00619 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.274444] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52080c9e-f15c-d5ab-afd6-2f0c94e45cef, 'name': SearchDatastore_Task, 'duration_secs': 0.012843} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.276044] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.276309] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.276575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.276731] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 812.276938] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.277348] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 812.277348] env[65788]: value = "task-4662362" [ 812.277348] env[65788]: _type = "Task" [ 812.277348] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.277654] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-027a7a6e-ed5b-4fda-9df1-5f67c811f619 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.293011] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662362, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.297146] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.297146] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.297938] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77b9f26b-acef-4cc3-a3cd-64e0f3ae50a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.304424] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 812.304424] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218ee05-24ae-52f4-e1d3-4a7892f60160" [ 812.304424] env[65788]: _type = "Task" [ 812.304424] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.319729] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218ee05-24ae-52f4-e1d3-4a7892f60160, 'name': SearchDatastore_Task, 'duration_secs': 0.01021} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.323618] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662361, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457978} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.323798] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54846a3f-d5ee-4918-9238-de80e7014392 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.326359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] caceb296-5da2-4b4d-b955-5d7a238ff939/caceb296-5da2-4b4d-b955-5d7a238ff939.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.326584] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.326842] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b31374d2-856c-4180-866e-773fe9b3ebba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.333155] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 812.333155] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525cc438-4d81-7bbf-da6b-ae81d34a0c0e" [ 812.333155] env[65788]: _type = "Task" [ 812.333155] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.337750] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 812.337750] env[65788]: value = "task-4662363" [ 812.337750] env[65788]: _type = "Task" [ 812.337750] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.347997] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525cc438-4d81-7bbf-da6b-ae81d34a0c0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.355884] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662363, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.540721] env[65788]: DEBUG nova.compute.manager [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65788) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:925}} [ 812.540990] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.800441] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662362, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07592} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.803994] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.805211] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0530513-b4b2-43eb-a016-4d070874ae5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.831948] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.835916] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d1c665f-4062-406b-8c1d-fc9ed8cd2f36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.866128] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662363, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071416} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.871047] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.871437] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525cc438-4d81-7bbf-da6b-ae81d34a0c0e, 'name': SearchDatastore_Task, 'duration_secs': 0.011506} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.871775] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 812.871775] env[65788]: value = "task-4662364" [ 812.871775] env[65788]: _type = "Task" [ 812.871775] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.875780] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0ddcf9-981d-4406-8d6b-1eb2b8aa83e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.878450] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 812.878450] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.878945] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ed5e318-2092-48e0-88c5-56c89984f1e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.914665] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] caceb296-5da2-4b4d-b955-5d7a238ff939/caceb296-5da2-4b4d-b955-5d7a238ff939.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.924131] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82e20699-9a7c-498c-a577-48ba35a6ba42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.942322] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.943067] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 812.943067] env[65788]: value = "task-4662365" [ 812.943067] env[65788]: _type = "Task" [ 812.943067] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.951233] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 812.951233] env[65788]: value = "task-4662366" [ 812.951233] env[65788]: _type = "Task" [ 812.951233] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.955283] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.968773] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.995432] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8f2f30-847f-48b4-b940-5524db3a67a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.005697] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be820e62-1161-4587-af0e-e56a1c102ddf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.044657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e095b9b-b49c-4d94-b237-9b531d36df14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.055224] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19edf83-aa02-4794-8fc2-4c21e752329c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.070943] env[65788]: DEBUG nova.compute.provider_tree [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.403569] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662364, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.459623] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55553} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.470225] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.470225] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.470225] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-694d8d4d-1d15-4863-8122-e23b81df2550 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.477020] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662366, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.478444] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 813.478444] env[65788]: value = "task-4662367" [ 813.478444] env[65788]: _type = "Task" [ 813.478444] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.491127] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662367, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.575401] env[65788]: DEBUG nova.scheduler.client.report [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 813.889568] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662364, 'name': ReconfigVM_Task, 'duration_secs': 0.541658} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.889863] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Reconfigured VM instance instance-00000032 to attach disk [datastore2] a813d102-1b77-4214-8eab-2cd66e002912/a813d102-1b77-4214-8eab-2cd66e002912.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.890649] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e9c9780-d37e-404d-a5ab-e1388145c169 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.900087] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 813.900087] env[65788]: value = "task-4662368" [ 813.900087] env[65788]: _type = "Task" [ 813.900087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.910492] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662368, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.974240] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662366, 'name': ReconfigVM_Task, 'duration_secs': 0.731124} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.974649] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Reconfigured VM instance instance-00000035 to attach disk [datastore2] caceb296-5da2-4b4d-b955-5d7a238ff939/caceb296-5da2-4b4d-b955-5d7a238ff939.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.976205] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-da01fded-85ab-461a-8476-8511c46a65ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.984543] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 813.984543] env[65788]: value = "task-4662369" [ 813.984543] env[65788]: _type = "Task" [ 813.984543] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.993674] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662367, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071091} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.994411] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.995278] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a246c2e7-0212-4e86-be4e-f4d5ba1bc0a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.001558] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662369, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.022628] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.023064] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d6b54ef-cef0-4584-a7dd-6066a8517538 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.046678] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 814.046678] env[65788]: value = "task-4662370" [ 814.046678] env[65788]: _type = "Task" [ 814.046678] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.057025] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.084411] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.757s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.085240] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 814.088836] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 43.518s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.089352] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.089352] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 814.089693] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 42.238s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.091271] env[65788]: INFO nova.compute.claims [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.095366] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8502e618-5c56-4287-8e7f-c8b7d0cc9f1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.107159] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e464ede-7c85-42e5-845e-a5456534fd91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.123515] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb2f456-deac-4484-be5d-db76e3df12e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.133961] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5d205c-7815-4c4e-bb24-b3c7b7b965f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.170426] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177821MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 814.170604] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.415320] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662368, 'name': Rename_Task, 'duration_secs': 0.203746} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.415829] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.416281] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01d9568d-c4fe-4156-9de6-a7e01c5745b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.426222] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Waiting for the task: (returnval){ [ 814.426222] env[65788]: value = "task-4662371" [ 814.426222] env[65788]: _type = "Task" [ 814.426222] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.441306] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662371, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.498876] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662369, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.557074] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.599031] env[65788]: DEBUG nova.compute.utils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 814.599031] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 814.606020] env[65788]: DEBUG nova.network.neutron [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 814.606020] env[65788]: WARNING neutronclient.v2_0.client [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.606020] env[65788]: WARNING neutronclient.v2_0.client [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 814.606020] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.606280] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.811283] env[65788]: DEBUG nova.policy [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48473ea517f64d3ba0051de1e9e378b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6dcc0ac928ee42b9a926c98464c4ec95', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 814.938778] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662371, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.000653] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662369, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.059148] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662370, 'name': ReconfigVM_Task, 'duration_secs': 0.72103} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.059848] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.060695] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1cc010b-d08c-4ade-aa43-50327afe8417 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.070578] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 815.070578] env[65788]: value = "task-4662372" [ 815.070578] env[65788]: _type = "Task" [ 815.070578] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.082735] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662372, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.116896] env[65788]: DEBUG nova.compute.utils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 815.309805] env[65788]: DEBUG nova.network.neutron [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Successfully created port: 773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 815.439795] env[65788]: DEBUG oslo_vmware.api [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Task: {'id': task-4662371, 'name': PowerOnVM_Task, 'duration_secs': 0.808998} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.443089] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 815.443607] env[65788]: DEBUG nova.compute.manager [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 815.444718] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6202fdb4-1368-47d6-aa6d-68de14989ec9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.507516] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662369, 'name': Rename_Task, 'duration_secs': 1.17973} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.507516] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.507516] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-629c9fee-16a9-4963-83c9-223bb20a8b65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.518560] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 815.518560] env[65788]: value = "task-4662373" [ 815.518560] env[65788]: _type = "Task" [ 815.518560] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.530192] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.585592] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662372, 'name': Rename_Task, 'duration_secs': 0.198981} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.585592] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.586608] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-443b25bb-439b-4fc5-a13e-14a5478f4bec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.594407] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 815.594407] env[65788]: value = "task-4662374" [ 815.594407] env[65788]: _type = "Task" [ 815.594407] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.608574] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.621199] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 815.691525] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803e1e9b-4d14-4c40-9212-517e4505bf6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.701251] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4c6617-74b7-49e6-9994-97599ddf362a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.744736] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded73459-482d-4196-878c-39805f521ee6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.755925] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23b7041-07f6-41f3-9d78-4d45e6bdb93f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.775197] env[65788]: DEBUG nova.compute.provider_tree [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.965410] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.030627] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662373, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.107058] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662374, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.280250] env[65788]: DEBUG nova.scheduler.client.report [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 816.339744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.340053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.533185] env[65788]: DEBUG oslo_vmware.api [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662373, 'name': PowerOnVM_Task, 'duration_secs': 0.523541} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.533560] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.533851] env[65788]: INFO nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Took 9.29 seconds to spawn the instance on the hypervisor. [ 816.534050] env[65788]: DEBUG nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 816.534988] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78120167-a880-418e-a120-092cd5f6fe80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.605913] env[65788]: DEBUG oslo_vmware.api [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662374, 'name': PowerOnVM_Task, 'duration_secs': 0.75184} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.606197] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.606389] env[65788]: INFO nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Took 5.51 seconds to spawn the instance on the hypervisor. [ 816.606559] env[65788]: DEBUG nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 816.607361] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d1f6cc-5fb3-4bea-9cc7-d0b847ae3dce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.632037] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 816.663869] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:45:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='1532738474',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-317301314',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 816.664245] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 816.664480] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 816.664758] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 816.665119] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 816.665223] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 816.665530] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.665777] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 816.666053] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 816.666310] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 816.666577] env[65788]: DEBUG nova.virt.hardware [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 816.668182] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fbf646-a5fa-4a32-bcca-0e1b33a372c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.679744] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547f6274-a86f-4dfa-88b1-3ab9498877ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.788234] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.698s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.788996] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 816.797216] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 44.260s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.797216] env[65788]: DEBUG nova.objects.instance [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 816.799868] env[65788]: DEBUG nova.compute.manager [req-f601036a-bf78-46dc-865f-77e043856f5d req-3cc01357-2129-4f97-8d4b-bc892d4cc343 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Received event network-vif-plugged-773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 816.800184] env[65788]: DEBUG oslo_concurrency.lockutils [req-f601036a-bf78-46dc-865f-77e043856f5d req-3cc01357-2129-4f97-8d4b-bc892d4cc343 service nova] Acquiring lock "bbf6a242-9873-44b9-8938-2f8df11a1018-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.800481] env[65788]: DEBUG oslo_concurrency.lockutils [req-f601036a-bf78-46dc-865f-77e043856f5d req-3cc01357-2129-4f97-8d4b-bc892d4cc343 service nova] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.800729] env[65788]: DEBUG oslo_concurrency.lockutils [req-f601036a-bf78-46dc-865f-77e043856f5d req-3cc01357-2129-4f97-8d4b-bc892d4cc343 service nova] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.800975] env[65788]: DEBUG nova.compute.manager [req-f601036a-bf78-46dc-865f-77e043856f5d req-3cc01357-2129-4f97-8d4b-bc892d4cc343 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] No waiting events found dispatching network-vif-plugged-773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 816.801228] env[65788]: WARNING nova.compute.manager [req-f601036a-bf78-46dc-865f-77e043856f5d req-3cc01357-2129-4f97-8d4b-bc892d4cc343 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Received unexpected event network-vif-plugged-773392a6-7451-4cad-95dd-aba2407f0497 for instance with vm_state building and task_state spawning. [ 816.843222] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 816.895049] env[65788]: DEBUG nova.network.neutron [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Successfully updated port: 773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 816.930434] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "a813d102-1b77-4214-8eab-2cd66e002912" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.930725] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "a813d102-1b77-4214-8eab-2cd66e002912" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.931248] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "a813d102-1b77-4214-8eab-2cd66e002912-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.931248] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "a813d102-1b77-4214-8eab-2cd66e002912-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.931446] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "a813d102-1b77-4214-8eab-2cd66e002912-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.935505] env[65788]: INFO nova.compute.manager [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Terminating instance [ 817.057265] env[65788]: INFO nova.compute.manager [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Took 56.28 seconds to build instance. [ 817.124033] env[65788]: INFO nova.compute.manager [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Took 50.37 seconds to build instance. [ 817.311219] env[65788]: DEBUG nova.compute.utils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 817.312608] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 817.312987] env[65788]: DEBUG nova.network.neutron [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 817.315185] env[65788]: WARNING neutronclient.v2_0.client [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.315185] env[65788]: WARNING neutronclient.v2_0.client [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.315884] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.316750] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.325985] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 817.377457] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.397535] env[65788]: DEBUG nova.policy [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c4a00121e214dc9a7b0866c8deee18e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffbe44d63c1d432e97849f15615329e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 817.399802] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.399960] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 817.400255] env[65788]: DEBUG nova.network.neutron [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 817.443872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "refresh_cache-a813d102-1b77-4214-8eab-2cd66e002912" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.444539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquired lock "refresh_cache-a813d102-1b77-4214-8eab-2cd66e002912" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 817.444539] env[65788]: DEBUG nova.network.neutron [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 817.559422] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5a9e8e0-de6b-48b7-9bdb-b41f328486d2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.945s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.626232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31690447-2243-4216-a328-ea8bd920dea0 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.605s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.764945] env[65788]: DEBUG nova.network.neutron [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Successfully created port: 31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 817.812915] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f56b82-b60b-7bcf-e003-6613bc07df66/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 817.815238] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d436e9c2-56d3-42ab-80aa-d79dac202233 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.817198] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67580dc4-8583-4814-b70a-ddd264f9bfe0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.822533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 44.491s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.823044] env[65788]: DEBUG nova.objects.instance [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lazy-loading 'resources' on Instance uuid 54ad3472-0cca-4dca-91f2-8c343f706926 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.838869] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f56b82-b60b-7bcf-e003-6613bc07df66/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 817.839090] env[65788]: ERROR oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f56b82-b60b-7bcf-e003-6613bc07df66/disk-0.vmdk due to incomplete transfer. [ 817.840224] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "caceb296-5da2-4b4d-b955-5d7a238ff939" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.840457] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.840769] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "caceb296-5da2-4b4d-b955-5d7a238ff939-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.841119] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.841216] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.842892] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-55840c55-6d96-4e61-ae0c-af4cbb07667d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.845065] env[65788]: INFO nova.compute.manager [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Terminating instance [ 817.855658] env[65788]: DEBUG oslo_vmware.rw_handles [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f56b82-b60b-7bcf-e003-6613bc07df66/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 817.856550] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Uploaded image 2ab5928d-b287-4fb9-8e31-7911bdbeafc7 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 817.859141] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 817.860370] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e5154ab8-09ca-4eee-99ee-a370c75dc4f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.873069] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 817.873069] env[65788]: value = "task-4662375" [ 817.873069] env[65788]: _type = "Task" [ 817.873069] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.886407] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662375, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.906860] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.910019] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.947860] env[65788]: WARNING neutronclient.v2_0.client [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 817.948784] env[65788]: WARNING openstack [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.949237] env[65788]: WARNING openstack [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.966621] env[65788]: DEBUG nova.network.neutron [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 817.994050] env[65788]: DEBUG nova.network.neutron [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 818.009791] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.010222] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.121209] env[65788]: DEBUG nova.network.neutron [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.165444] env[65788]: WARNING neutronclient.v2_0.client [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 818.166225] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 818.166625] env[65788]: WARNING openstack [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 818.348059] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 818.351141] env[65788]: DEBUG nova.compute.manager [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 818.353083] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.355444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b507419-a513-4fe2-962d-1667ea3bf9a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.366777] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.367249] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b40eaf32-cc71-4749-bbee-a33d11d6d50e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.376827] env[65788]: DEBUG nova.network.neutron [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Updating instance_info_cache with network_info: [{"id": "773392a6-7451-4cad-95dd-aba2407f0497", "address": "fa:16:3e:04:ba:6b", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773392a6-74", "ovs_interfaceid": "773392a6-7451-4cad-95dd-aba2407f0497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 818.385364] env[65788]: DEBUG oslo_vmware.api [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 818.385364] env[65788]: value = "task-4662376" [ 818.385364] env[65788]: _type = "Task" [ 818.385364] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.399114] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 818.399476] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 818.399781] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 818.400124] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 818.400370] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 818.400580] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 818.400891] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.401146] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 818.401373] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 818.401621] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 818.401858] env[65788]: DEBUG nova.virt.hardware [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 818.402271] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662375, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.403217] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202ff02c-349c-4938-95fa-514860ec6248 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.417088] env[65788]: DEBUG oslo_vmware.api [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662376, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.420987] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240b558d-fba8-4e6a-811d-e761ea380c47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.626972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Releasing lock "refresh_cache-a813d102-1b77-4214-8eab-2cd66e002912" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.627252] env[65788]: DEBUG nova.compute.manager [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 818.627489] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.630705] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43420344-1ff0-4f81-adb1-dd2f951890bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.638686] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.638981] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98c6cd13-5b2d-42ec-a7c3-4a1792ad0bc0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.646530] env[65788]: DEBUG oslo_vmware.api [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 818.646530] env[65788]: value = "task-4662377" [ 818.646530] env[65788]: _type = "Task" [ 818.646530] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.659555] env[65788]: DEBUG oslo_vmware.api [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662377, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.877697] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84061ade-3e0f-43d3-b98b-27b526fd37a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.891712] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Releasing lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.892111] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Instance network_info: |[{"id": "773392a6-7451-4cad-95dd-aba2407f0497", "address": "fa:16:3e:04:ba:6b", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773392a6-74", "ovs_interfaceid": "773392a6-7451-4cad-95dd-aba2407f0497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 818.892590] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662375, 'name': Destroy_Task, 'duration_secs': 0.65145} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.896079] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:ba:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4a8f5363-be3a-4f92-9ccf-33bb0c8113b3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '773392a6-7451-4cad-95dd-aba2407f0497', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.904105] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 818.907545] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Destroyed the VM [ 818.908194] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 818.908289] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.908482] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-479933d1-129e-46fc-af46-14dd7412fb0d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.911485] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a981a37-cb09-4ea1-bd3f-d58a067155cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.914733] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89812dfb-354a-4279-9109-ac965e0b2fc4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.943384] env[65788]: DEBUG oslo_vmware.api [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662376, 'name': PowerOffVM_Task, 'duration_secs': 0.236289} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.943688] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 818.943688] env[65788]: value = "task-4662378" [ 818.943688] env[65788]: _type = "Task" [ 818.943688] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.970896] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 818.971135] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 818.971761] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.971761] env[65788]: value = "task-4662379" [ 818.971761] env[65788]: _type = "Task" [ 818.971761] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.973183] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da7cfee8-4f1f-4e9c-a7d1-531ff5338472 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.978481] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2243b20-8b76-4995-a8dc-a4aa29c96412 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.008813] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384c41e6-d2a6-4264-8821-e38d2336f914 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.013679] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662378, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.014939] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662379, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.028314] env[65788]: DEBUG nova.compute.provider_tree [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.066392] env[65788]: DEBUG nova.compute.manager [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Received event network-changed-773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 819.066693] env[65788]: DEBUG nova.compute.manager [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Refreshing instance network info cache due to event network-changed-773392a6-7451-4cad-95dd-aba2407f0497. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 819.067366] env[65788]: DEBUG oslo_concurrency.lockutils [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Acquiring lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.067366] env[65788]: DEBUG oslo_concurrency.lockutils [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Acquired lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.067366] env[65788]: DEBUG nova.network.neutron [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Refreshing network info cache for port 773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 819.074363] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 819.074363] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 819.074363] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleting the datastore file [datastore2] caceb296-5da2-4b4d-b955-5d7a238ff939 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.074363] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ed7bf21-6647-4586-88a7-5dfc7a1f4c30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.081513] env[65788]: DEBUG oslo_vmware.api [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 819.081513] env[65788]: value = "task-4662381" [ 819.081513] env[65788]: _type = "Task" [ 819.081513] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.092578] env[65788]: DEBUG oslo_vmware.api [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662381, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.159348] env[65788]: DEBUG oslo_vmware.api [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662377, 'name': PowerOffVM_Task, 'duration_secs': 0.244345} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.159726] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.159911] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 819.160262] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1daaafb4-c122-410d-a55b-95664044b56e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.192022] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 819.192144] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 819.192344] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Deleting the datastore file [datastore2] a813d102-1b77-4214-8eab-2cd66e002912 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 819.192748] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23198031-042a-4b49-95f7-bfc66b11c94e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.202356] env[65788]: DEBUG oslo_vmware.api [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for the task: (returnval){ [ 819.202356] env[65788]: value = "task-4662383" [ 819.202356] env[65788]: _type = "Task" [ 819.202356] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.213777] env[65788]: DEBUG oslo_vmware.api [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.388776] env[65788]: DEBUG nova.network.neutron [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Successfully updated port: 31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 819.481386] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662378, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.489917] env[65788]: INFO nova.compute.manager [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Rebuilding instance [ 819.491968] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662379, 'name': CreateVM_Task, 'duration_secs': 0.381273} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.492344] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.492851] env[65788]: WARNING neutronclient.v2_0.client [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.493235] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.493385] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.493694] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 819.493965] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74f05e2b-56df-4927-b96e-1e3386901cee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.499086] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 819.499086] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52230795-455b-343a-3750-130ac2a37a38" [ 819.499086] env[65788]: _type = "Task" [ 819.499086] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.508470] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52230795-455b-343a-3750-130ac2a37a38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.531723] env[65788]: DEBUG nova.scheduler.client.report [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 819.542469] env[65788]: DEBUG nova.compute.manager [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 819.543389] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eecf717-8446-4c21-83ed-4a0a968c7b56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.573946] env[65788]: WARNING neutronclient.v2_0.client [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.574748] env[65788]: WARNING openstack [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.575173] env[65788]: WARNING openstack [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.601313] env[65788]: DEBUG oslo_vmware.api [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662381, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176369} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.602062] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.602062] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 819.602062] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.602198] env[65788]: INFO nova.compute.manager [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Took 1.25 seconds to destroy the instance on the hypervisor. [ 819.602355] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 819.602550] env[65788]: DEBUG nova.compute.manager [-] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 819.602650] env[65788]: DEBUG nova.network.neutron [-] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 819.603079] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.603460] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.603662] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.643943] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.686983] env[65788]: WARNING openstack [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.687236] env[65788]: WARNING openstack [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.712838] env[65788]: DEBUG oslo_vmware.api [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Task: {'id': task-4662383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146157} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.713052] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 819.713244] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 819.713418] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 819.713593] env[65788]: INFO nova.compute.manager [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Took 1.09 seconds to destroy the instance on the hypervisor. [ 819.713847] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 819.714056] env[65788]: DEBUG nova.compute.manager [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 819.714156] env[65788]: DEBUG nova.network.neutron [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 819.714405] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.714940] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.717030] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.738607] env[65788]: DEBUG nova.network.neutron [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 819.738860] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.775749] env[65788]: WARNING neutronclient.v2_0.client [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 819.776630] env[65788]: WARNING openstack [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.777406] env[65788]: WARNING openstack [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.868685] env[65788]: DEBUG nova.network.neutron [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Updated VIF entry in instance network info cache for port 773392a6-7451-4cad-95dd-aba2407f0497. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 819.869054] env[65788]: DEBUG nova.network.neutron [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Updating instance_info_cache with network_info: [{"id": "773392a6-7451-4cad-95dd-aba2407f0497", "address": "fa:16:3e:04:ba:6b", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773392a6-74", "ovs_interfaceid": "773392a6-7451-4cad-95dd-aba2407f0497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.894394] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.894394] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.894394] env[65788]: DEBUG nova.network.neutron [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 819.912096] env[65788]: DEBUG nova.compute.manager [req-b27bfe01-4569-4f08-9e5e-2d43515f089b req-dcedf896-a9a0-448f-928f-96c63874f08e service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Received event network-vif-deleted-29e1dcd1-5465-47b7-929e-de731202dbd0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 819.912400] env[65788]: INFO nova.compute.manager [req-b27bfe01-4569-4f08-9e5e-2d43515f089b req-dcedf896-a9a0-448f-928f-96c63874f08e service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Neutron deleted interface 29e1dcd1-5465-47b7-929e-de731202dbd0; detaching it from the instance and deleting it from the info cache [ 819.912638] env[65788]: DEBUG nova.network.neutron [req-b27bfe01-4569-4f08-9e5e-2d43515f089b req-dcedf896-a9a0-448f-928f-96c63874f08e service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 819.984700] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662378, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.009918] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52230795-455b-343a-3750-130ac2a37a38, 'name': SearchDatastore_Task, 'duration_secs': 0.01056} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.010258] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.010490] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 820.010722] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.010896] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 820.011101] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 820.011372] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad076a40-ae62-4d1c-9478-19e3878aea72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.021654] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 820.021906] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 820.022723] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be4acb65-f79f-4863-9d5c-2b8507dd91b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.029237] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 820.029237] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525928de-20f1-8a43-5eef-4788d2025b5a" [ 820.029237] env[65788]: _type = "Task" [ 820.029237] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.039099] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.216s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.041354] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525928de-20f1-8a43-5eef-4788d2025b5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.041872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 46.499s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.042138] env[65788]: DEBUG nova.objects.instance [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lazy-loading 'resources' on Instance uuid 99f2cc13-a559-4379-8332-d432e8bac36b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.068238] env[65788]: INFO nova.scheduler.client.report [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Deleted allocations for instance 54ad3472-0cca-4dca-91f2-8c343f706926 [ 820.241113] env[65788]: DEBUG nova.network.neutron [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 820.367796] env[65788]: DEBUG nova.network.neutron [-] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 820.372123] env[65788]: DEBUG oslo_concurrency.lockutils [req-51a643ec-5b06-4288-94e3-4e242421d397 req-b5f69054-11e9-4c31-9574-b51f2bf30fca service nova] Releasing lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.398936] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.399367] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.415278] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58b478ec-663e-4a8c-b6b6-a8e9ca208666 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.425884] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9174ec3-d49d-43ea-bb5b-ffb9d037ef4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.441016] env[65788]: DEBUG nova.network.neutron [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 820.465279] env[65788]: DEBUG nova.compute.manager [req-b27bfe01-4569-4f08-9e5e-2d43515f089b req-dcedf896-a9a0-448f-928f-96c63874f08e service nova] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Detach interface failed, port_id=29e1dcd1-5465-47b7-929e-de731202dbd0, reason: Instance caceb296-5da2-4b4d-b955-5d7a238ff939 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 820.466424] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.466779] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.482359] env[65788]: DEBUG oslo_vmware.api [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662378, 'name': RemoveSnapshot_Task, 'duration_secs': 1.199009} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.482670] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 820.482904] env[65788]: INFO nova.compute.manager [None req-62cc1b33-6833-46e9-878f-c01a7979b1b2 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Took 18.45 seconds to snapshot the instance on the hypervisor. [ 820.542873] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525928de-20f1-8a43-5eef-4788d2025b5a, 'name': SearchDatastore_Task, 'duration_secs': 0.010162} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.543678] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e33388af-83fc-40af-94a6-8350248430be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.551429] env[65788]: WARNING neutronclient.v2_0.client [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 820.552101] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.552467] env[65788]: WARNING openstack [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.561559] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.563810] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f9a325e-1d0f-4f7f-9a59-3b7927edcbbc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.565541] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 820.565541] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5245044a-ea69-ecf6-a280-1803a26cf2ec" [ 820.565541] env[65788]: _type = "Task" [ 820.565541] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.576519] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 820.576519] env[65788]: value = "task-4662384" [ 820.576519] env[65788]: _type = "Task" [ 820.576519] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.580559] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5245044a-ea69-ecf6-a280-1803a26cf2ec, 'name': SearchDatastore_Task, 'duration_secs': 0.010623} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.581074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-df19d71d-b322-4b8b-a74e-4efdfcb3b970 tempest-ServersTestBootFromVolume-1696516378 tempest-ServersTestBootFromVolume-1696516378-project-member] Lock "54ad3472-0cca-4dca-91f2-8c343f706926" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 52.862s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.585125] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.585410] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018/bbf6a242-9873-44b9-8938-2f8df11a1018.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.588362] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77584232-55be-4527-9b2d-a1ef8d28b9d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.598071] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662384, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.599600] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 820.599600] env[65788]: value = "task-4662385" [ 820.599600] env[65788]: _type = "Task" [ 820.599600] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.617750] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662385, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.652245] env[65788]: DEBUG nova.network.neutron [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updating instance_info_cache with network_info: [{"id": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "address": "fa:16:3e:91:8c:07", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31daf1d5-0a", "ovs_interfaceid": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 820.744015] env[65788]: INFO nova.compute.manager [-] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Took 1.03 seconds to deallocate network for instance. [ 820.871329] env[65788]: INFO nova.compute.manager [-] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Took 1.27 seconds to deallocate network for instance. [ 821.052870] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b3a8e2-4ca5-4d72-82a5-5e135ef1bd41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.062268] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882ba7a3-a95a-44f6-9d8f-0df933811a68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.100545] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a708a3-95df-4c82-9f5f-1b3204cb6392 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.105509] env[65788]: DEBUG nova.compute.manager [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Received event network-vif-plugged-31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 821.105769] env[65788]: DEBUG oslo_concurrency.lockutils [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.106130] env[65788]: DEBUG oslo_concurrency.lockutils [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.106330] env[65788]: DEBUG oslo_concurrency.lockutils [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.106500] env[65788]: DEBUG nova.compute.manager [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] No waiting events found dispatching network-vif-plugged-31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 821.106717] env[65788]: WARNING nova.compute.manager [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Received unexpected event network-vif-plugged-31daf1d5-0a84-4d17-a694-a7ab16e9b81b for instance with vm_state building and task_state spawning. [ 821.106885] env[65788]: DEBUG nova.compute.manager [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Received event network-changed-31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 821.107045] env[65788]: DEBUG nova.compute.manager [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Refreshing instance network info cache due to event network-changed-31daf1d5-0a84-4d17-a694-a7ab16e9b81b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 821.107229] env[65788]: DEBUG oslo_concurrency.lockutils [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Acquiring lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.121998] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09256bf-3a7a-416a-b88c-f44c13e03206 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.126298] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662385, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511323} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.126823] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662384, 'name': PowerOffVM_Task, 'duration_secs': 0.192448} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.127071] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018/bbf6a242-9873-44b9-8938-2f8df11a1018.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.127278] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.127525] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 821.128243] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.128838] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e01462d4-ea16-4bba-9c7d-327373bf31e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.131654] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6402f870-8752-479d-8d1f-c88b95077239 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.143301] env[65788]: DEBUG nova.compute.provider_tree [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.151656] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 821.153069] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf26aedc-462b-4f9a-9f06-edff0b917806 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.156165] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 821.156165] env[65788]: value = "task-4662386" [ 821.156165] env[65788]: _type = "Task" [ 821.156165] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.156165] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 821.156358] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance network_info: |[{"id": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "address": "fa:16:3e:91:8c:07", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31daf1d5-0a", "ovs_interfaceid": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 821.156488] env[65788]: DEBUG oslo_concurrency.lockutils [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Acquired lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.156675] env[65788]: DEBUG nova.network.neutron [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Refreshing network info cache for port 31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 821.158060] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:8c:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31daf1d5-0a84-4d17-a694-a7ab16e9b81b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.165526] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 821.170656] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.171227] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dfe84379-edb1-4738-95d5-d92bc95973e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.197946] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.200048] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.200048] env[65788]: value = "task-4662388" [ 821.200048] env[65788]: _type = "Task" [ 821.200048] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.207178] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 821.207495] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 821.207759] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Deleting the datastore file [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 821.208955] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2665897b-5386-475c-8a53-d840b64358e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.215297] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662388, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.222217] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 821.222217] env[65788]: value = "task-4662389" [ 821.222217] env[65788]: _type = "Task" [ 821.222217] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.231187] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.251870] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.379703] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.647196] env[65788]: DEBUG nova.scheduler.client.report [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 821.668589] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094373} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.669587] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 821.670237] env[65788]: WARNING neutronclient.v2_0.client [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.670876] env[65788]: WARNING openstack [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.671242] env[65788]: WARNING openstack [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.678889] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b84aca7-4147-497e-b7b1-805ecde3401d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.702982] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018/bbf6a242-9873-44b9-8938-2f8df11a1018.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 821.703705] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-921c9e91-9a56-45bd-8533-6e1ff1ba3131 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.731631] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 821.731631] env[65788]: value = "task-4662390" [ 821.731631] env[65788]: _type = "Task" [ 821.731631] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.737985] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176288} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.737985] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662388, 'name': CreateVM_Task, 'duration_secs': 0.444473} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.742217] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.742424] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 821.742934] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.745240] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.746164] env[65788]: WARNING neutronclient.v2_0.client [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.746564] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.746726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.747096] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 821.747976] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e39ceec-33e1-47dd-bdf6-067be9f4cf66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.759225] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.765665] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 821.765665] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c8570-6cde-d682-36a7-5bf36a3068d9" [ 821.765665] env[65788]: _type = "Task" [ 821.765665] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.773698] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c8570-6cde-d682-36a7-5bf36a3068d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.847917] env[65788]: WARNING openstack [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.847917] env[65788]: WARNING openstack [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.918417] env[65788]: WARNING neutronclient.v2_0.client [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 821.919063] env[65788]: WARNING openstack [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.919792] env[65788]: WARNING openstack [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.029960] env[65788]: DEBUG nova.network.neutron [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updated VIF entry in instance network info cache for port 31daf1d5-0a84-4d17-a694-a7ab16e9b81b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 822.030349] env[65788]: DEBUG nova.network.neutron [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updating instance_info_cache with network_info: [{"id": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "address": "fa:16:3e:91:8c:07", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31daf1d5-0a", "ovs_interfaceid": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 822.154137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.111s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.156950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 46.380s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.159242] env[65788]: INFO nova.compute.claims [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.192356] env[65788]: INFO nova.scheduler.client.report [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted allocations for instance 99f2cc13-a559-4379-8332-d432e8bac36b [ 822.256784] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662390, 'name': ReconfigVM_Task, 'duration_secs': 0.369262} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.258866] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Reconfigured VM instance instance-00000037 to attach disk [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018/bbf6a242-9873-44b9-8938-2f8df11a1018.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.259725] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=65788) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 822.261214] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-bfe54467-de97-4945-8508-f1fae8c4ae69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.279984] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 822.279984] env[65788]: value = "task-4662391" [ 822.279984] env[65788]: _type = "Task" [ 822.279984] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.287493] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c8570-6cde-d682-36a7-5bf36a3068d9, 'name': SearchDatastore_Task, 'duration_secs': 0.011463} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.292534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.292825] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.293130] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.293223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.293410] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.294708] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-267c83dc-c68f-423a-9c34-4377d94568e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.304089] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662391, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.307222] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.307222] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.308842] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae8903ef-4d1a-4997-a1eb-a48d9c33b145 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.316049] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 822.316049] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526314c9-f60a-a160-5e91-6437ac3aa30b" [ 822.316049] env[65788]: _type = "Task" [ 822.316049] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.328022] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526314c9-f60a-a160-5e91-6437ac3aa30b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.534038] env[65788]: DEBUG oslo_concurrency.lockutils [req-0d3f9bce-42a2-4bdc-bcd1-80fbe82a9afa req-523c7b82-312d-4b05-81e3-285f285abe16 service nova] Releasing lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.702562] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8da413eb-908b-43c8-a93b-3c403e2f3040 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "99f2cc13-a559-4379-8332-d432e8bac36b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 52.163s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.797290] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 822.797605] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 822.797709] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 822.797888] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 822.798123] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 822.798371] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 822.798589] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.798750] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 822.798923] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 822.799193] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 822.799271] env[65788]: DEBUG nova.virt.hardware [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 822.800208] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3752e2d3-ddfc-4b95-8afd-869aa1ffc7e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.811024] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662391, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.058604} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.813399] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=65788) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 822.814668] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d69e773-b1cb-48ba-a618-b30c694dc84e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.818376] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2ddd4c-e337-4168-8a3e-933cf3ce9916 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.833628] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526314c9-f60a-a160-5e91-6437ac3aa30b, 'name': SearchDatastore_Task, 'duration_secs': 0.011088} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.863113] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018/ephemeral_0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.863659] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.869255] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 822.869569] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bb2bf14-c15d-4f1d-a575-24e8bf526176 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.872123] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d22a4467-589d-4ecb-87da-49710f1399df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.886258] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.886258] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-35a2ff02-99c0-4029-91a2-3220a3c054b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.903114] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 822.903114] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52acc9b1-51e0-aa9b-0c74-11503cc7314e" [ 822.903114] env[65788]: _type = "Task" [ 822.903114] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.909098] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.909098] env[65788]: value = "task-4662392" [ 822.909098] env[65788]: _type = "Task" [ 822.909098] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.909417] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 822.909417] env[65788]: value = "task-4662393" [ 822.909417] env[65788]: _type = "Task" [ 822.909417] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.919805] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52acc9b1-51e0-aa9b-0c74-11503cc7314e, 'name': SearchDatastore_Task, 'duration_secs': 0.010892} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.920946] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.921259] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 822.921595] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-379d91e2-2316-4ebe-8a49-c7575ea2d595 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.931328] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662393, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.931628] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662392, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.938724] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 822.938724] env[65788]: value = "task-4662394" [ 822.938724] env[65788]: _type = "Task" [ 822.938724] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.954091] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662394, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.287099] env[65788]: DEBUG nova.compute.manager [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 823.288179] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7a5dcb-d7fa-450c-a86e-c49d8bc6cb05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.432620] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662392, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.432620] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662393, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.452825] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662394, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.730978] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3518d1f5-5c82-432f-828a-c7e909dbf525 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.739647] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3e68e1-8b2a-4972-9188-559557d876a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.773600] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5e2a30-10b1-441d-934e-a6fd5264be79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.782343] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3293b09-b80a-4d2c-94a5-baa3a8ce3905 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.798779] env[65788]: DEBUG nova.compute.provider_tree [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.801902] env[65788]: INFO nova.compute.manager [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] instance snapshotting [ 823.804421] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02ee3e8-4548-490d-9733-a8c60a13e947 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.824766] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273a3662-24b9-48f4-aece-38e21bbc6288 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.926535] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662392, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.929954] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662393, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.951370] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662394, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545747} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.951723] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 823.952026] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 823.952305] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f29c11a-6a52-4975-99d5-81f87ecf74be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.960332] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 823.960332] env[65788]: value = "task-4662395" [ 823.960332] env[65788]: _type = "Task" [ 823.960332] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.970074] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662395, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.305296] env[65788]: DEBUG nova.scheduler.client.report [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.342019] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 824.342019] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4adf66e7-1a3a-4e1f-91b9-0b5f13a29dff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.351013] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 824.351013] env[65788]: value = "task-4662396" [ 824.351013] env[65788]: _type = "Task" [ 824.351013] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.360874] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662396, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.425179] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662392, 'name': CreateVM_Task, 'duration_secs': 1.3292} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.427823] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.428125] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662393, 'name': ReconfigVM_Task, 'duration_secs': 1.368761} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.428505] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.428658] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.429033] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 824.429318] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Reconfigured VM instance instance-00000037 to attach disk [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018/ephemeral_0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.429933] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14d4834c-fc16-412a-b3e2-28f927273db3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.431583] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e6fd93c-5e05-4623-b935-a286304ae588 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.437154] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 824.437154] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521d75da-054a-eb00-49cc-c07d5e179ce5" [ 824.437154] env[65788]: _type = "Task" [ 824.437154] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.438436] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 824.438436] env[65788]: value = "task-4662397" [ 824.438436] env[65788]: _type = "Task" [ 824.438436] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.450436] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521d75da-054a-eb00-49cc-c07d5e179ce5, 'name': SearchDatastore_Task, 'duration_secs': 0.009985} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.453758] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.454079] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.454320] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.454466] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.454643] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.454978] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662397, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.455137] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6eef3c48-6fc2-4737-860a-7ebbb30d21d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.466415] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.466713] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 824.468018] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b54bd19e-ca92-4575-8a11-59275ba8cd24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.473500] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662395, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310958} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.474460] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.475102] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e97e126-0de2-4759-b4d3-971d4a058dd1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.481495] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 824.481495] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52919fb9-2ce8-641f-955c-866adacb3c41" [ 824.481495] env[65788]: _type = "Task" [ 824.481495] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.504967] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.506183] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-414aeeee-c095-4cae-b08c-1c0343fcf4f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.524445] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52919fb9-2ce8-641f-955c-866adacb3c41, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.525720] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8ecc411-5e2b-4ee3-93ef-2a22b5c21284 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.534038] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 824.534038] env[65788]: value = "task-4662398" [ 824.534038] env[65788]: _type = "Task" [ 824.534038] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.534038] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 824.534038] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ef806-e41f-2586-906f-203f4775a90e" [ 824.534038] env[65788]: _type = "Task" [ 824.534038] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.548057] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ef806-e41f-2586-906f-203f4775a90e, 'name': SearchDatastore_Task, 'duration_secs': 0.010882} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.551201] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.551503] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.551875] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662398, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.552190] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d9742e26-faa8-4c4b-8314-3e614d1c776f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.564020] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 824.564020] env[65788]: value = "task-4662399" [ 824.564020] env[65788]: _type = "Task" [ 824.564020] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.570901] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662399, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.811380] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.811998] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 824.814816] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 46.849s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.866816] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662396, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.953809] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662397, 'name': Rename_Task, 'duration_secs': 0.1869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.954359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 824.958027] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de301e54-d69d-450d-b75c-26239238a6b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.963504] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 824.963504] env[65788]: value = "task-4662400" [ 824.963504] env[65788]: _type = "Task" [ 824.963504] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.975344] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.049586] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.072757] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662399, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.319807] env[65788]: DEBUG nova.compute.utils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 825.327554] env[65788]: INFO nova.compute.claims [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.339023] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 825.339023] env[65788]: DEBUG nova.network.neutron [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 825.339023] env[65788]: WARNING neutronclient.v2_0.client [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.339023] env[65788]: WARNING neutronclient.v2_0.client [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 825.339023] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 825.339444] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 825.365357] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662396, 'name': CreateSnapshot_Task, 'duration_secs': 0.868534} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.365933] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 825.366924] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f09111-2bf6-40d6-a1dd-e57bb88aed9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.403837] env[65788]: DEBUG nova.policy [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a1e3cda6bf9439982465aec80977cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1713c7002a413fb27180469fded83e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 825.476714] env[65788]: DEBUG oslo_vmware.api [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662400, 'name': PowerOnVM_Task, 'duration_secs': 0.497895} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.477213] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.477474] env[65788]: INFO nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Took 8.85 seconds to spawn the instance on the hypervisor. [ 825.478014] env[65788]: DEBUG nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 825.478563] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddcb5f5e-6a98-47d2-8b3f-34abedb31594 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.544521] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662398, 'name': ReconfigVM_Task, 'duration_secs': 0.65172} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.545153] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfigured VM instance instance-00000038 to attach disk [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.545930] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5dedbe1-1eb8-44cc-84a9-3613517ee0fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.553896] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 825.553896] env[65788]: value = "task-4662401" [ 825.553896] env[65788]: _type = "Task" [ 825.553896] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.563413] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662401, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.574558] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662399, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568548} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.574847] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.575092] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.575369] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d18e78f-41bc-4e77-b8fe-400d471563c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.585569] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 825.585569] env[65788]: value = "task-4662402" [ 825.585569] env[65788]: _type = "Task" [ 825.585569] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.596270] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662402, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.742160] env[65788]: DEBUG nova.network.neutron [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Successfully created port: ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 825.826078] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 825.837451] env[65788]: INFO nova.compute.resource_tracker [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating resource usage from migration b5064230-67b0-4b10-b1f0-9c2db8d69267 [ 825.893744] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 825.896878] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bd51d311-1552-4399-8513-f1847b8551db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.907981] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 825.907981] env[65788]: value = "task-4662403" [ 825.907981] env[65788]: _type = "Task" [ 825.907981] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.925383] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662403, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.000714] env[65788]: INFO nova.compute.manager [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Took 56.64 seconds to build instance. [ 826.066802] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662401, 'name': Rename_Task, 'duration_secs': 0.164319} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.067066] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.067412] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13b82929-cb0d-4037-83f9-3cf6135278a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.076695] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 826.076695] env[65788]: value = "task-4662404" [ 826.076695] env[65788]: _type = "Task" [ 826.076695] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.085940] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.098326] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662402, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078126} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.098605] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.099448] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06ceafe-2d47-4be6-9345-ffe850f0b9d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.128921] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.131944] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9e3d2fe-b247-4a07-a998-b68c78175344 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.155890] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 826.155890] env[65788]: value = "task-4662405" [ 826.155890] env[65788]: _type = "Task" [ 826.155890] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.169255] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662405, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.426139] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662403, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.434658] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064ce42a-3778-4fb9-95aa-773457934940 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.443544] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f692e8-d4ef-4633-b0e6-56e055ad9801 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.479432] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42045ab-a419-403e-b06e-87e685017513 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.490177] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984501bb-842e-4050-ac52-4718cec15b40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.505746] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f1c312a-22c1-4b20-a202-7bf7b0ec00df tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.483s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 826.506365] env[65788]: DEBUG nova.compute.provider_tree [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.589235] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662404, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.667568] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662405, 'name': ReconfigVM_Task, 'duration_secs': 0.333901} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.667879] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd/93e1ade8-4c57-4db9-9da3-388e3eea91bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.668591] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6767c4bd-65a9-46dc-aed5-c313a1ca463c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.676029] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 826.676029] env[65788]: value = "task-4662406" [ 826.676029] env[65788]: _type = "Task" [ 826.676029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.686244] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662406, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.838296] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 826.864560] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 826.864850] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 826.865304] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 826.865613] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 826.865828] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 826.865989] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 826.866258] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.866456] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 826.866833] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 826.866833] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 826.869441] env[65788]: DEBUG nova.virt.hardware [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 826.869441] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131af2d6-6be7-4404-a2a8-fa3c770e9c36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.879010] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b16d63-b930-4a39-a85c-d10e581869f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.921136] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662403, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.010483] env[65788]: DEBUG nova.scheduler.client.report [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 827.090880] env[65788]: DEBUG oslo_vmware.api [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662404, 'name': PowerOnVM_Task, 'duration_secs': 0.712083} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.091291] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.091396] env[65788]: INFO nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Took 8.74 seconds to spawn the instance on the hypervisor. [ 827.092057] env[65788]: DEBUG nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 827.093045] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8eeb58c-671b-40d4-b915-b97b2789bfee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.196696] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662406, 'name': Rename_Task, 'duration_secs': 0.181798} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.197347] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 827.198177] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0007d9f-768a-4e28-9b46-c1885ccae0ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.211543] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 827.211543] env[65788]: value = "task-4662407" [ 827.211543] env[65788]: _type = "Task" [ 827.211543] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.225060] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662407, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.426033] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662403, 'name': CloneVM_Task, 'duration_secs': 1.374542} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.426033] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Created linked-clone VM from snapshot [ 827.430587] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23160d7f-d520-4cfe-9879-d4418cf82c7c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.445075] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Uploading image aec2bed9-f33a-4c13-860a-a321f69ec40b {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 827.449744] env[65788]: DEBUG nova.network.neutron [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Successfully updated port: ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 827.485157] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 827.485157] env[65788]: value = "vm-910283" [ 827.485157] env[65788]: _type = "VirtualMachine" [ 827.485157] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 827.489019] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-67acc261-6a73-44e6-adc9-64091f521e34 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.495491] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lease: (returnval){ [ 827.495491] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b9f79-ea5c-e56b-9bcb-123229f179bd" [ 827.495491] env[65788]: _type = "HttpNfcLease" [ 827.495491] env[65788]: } obtained for exporting VM: (result){ [ 827.495491] env[65788]: value = "vm-910283" [ 827.495491] env[65788]: _type = "VirtualMachine" [ 827.495491] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 827.495812] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the lease: (returnval){ [ 827.495812] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b9f79-ea5c-e56b-9bcb-123229f179bd" [ 827.495812] env[65788]: _type = "HttpNfcLease" [ 827.495812] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 827.504287] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 827.504287] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b9f79-ea5c-e56b-9bcb-123229f179bd" [ 827.504287] env[65788]: _type = "HttpNfcLease" [ 827.504287] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 827.515844] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.701s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.516040] env[65788]: INFO nova.compute.manager [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Migrating [ 827.523891] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 46.588s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.524264] env[65788]: DEBUG nova.objects.instance [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lazy-loading 'resources' on Instance uuid 670f59c3-10b0-431f-a8b0-ef6f82a938be {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.592184] env[65788]: DEBUG nova.compute.manager [req-2f680c61-6dc6-4fe3-bc53-657a5fe186ff req-a5d6b016-3ea5-42e0-a06f-554997c6d444 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Received event network-vif-plugged-ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 827.592323] env[65788]: DEBUG oslo_concurrency.lockutils [req-2f680c61-6dc6-4fe3-bc53-657a5fe186ff req-a5d6b016-3ea5-42e0-a06f-554997c6d444 service nova] Acquiring lock "37471e59-1809-4df3-8c40-20921d04d18e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.592603] env[65788]: DEBUG oslo_concurrency.lockutils [req-2f680c61-6dc6-4fe3-bc53-657a5fe186ff req-a5d6b016-3ea5-42e0-a06f-554997c6d444 service nova] Lock "37471e59-1809-4df3-8c40-20921d04d18e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.592840] env[65788]: DEBUG oslo_concurrency.lockutils [req-2f680c61-6dc6-4fe3-bc53-657a5fe186ff req-a5d6b016-3ea5-42e0-a06f-554997c6d444 service nova] Lock "37471e59-1809-4df3-8c40-20921d04d18e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.593105] env[65788]: DEBUG nova.compute.manager [req-2f680c61-6dc6-4fe3-bc53-657a5fe186ff req-a5d6b016-3ea5-42e0-a06f-554997c6d444 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] No waiting events found dispatching network-vif-plugged-ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 827.593317] env[65788]: WARNING nova.compute.manager [req-2f680c61-6dc6-4fe3-bc53-657a5fe186ff req-a5d6b016-3ea5-42e0-a06f-554997c6d444 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Received unexpected event network-vif-plugged-ef83244d-23de-4589-bab0-ea354e648ac2 for instance with vm_state building and task_state spawning. [ 827.613717] env[65788]: INFO nova.compute.manager [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Took 55.80 seconds to build instance. [ 827.727949] env[65788]: DEBUG oslo_vmware.api [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662407, 'name': PowerOnVM_Task, 'duration_secs': 0.497894} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.728224] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.728415] env[65788]: DEBUG nova.compute.manager [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 827.729248] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e603952-e234-4bf8-af5c-b71609ed0a45 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.953832] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "refresh_cache-37471e59-1809-4df3-8c40-20921d04d18e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.954161] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "refresh_cache-37471e59-1809-4df3-8c40-20921d04d18e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 827.954379] env[65788]: DEBUG nova.network.neutron [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 828.007417] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 828.007417] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b9f79-ea5c-e56b-9bcb-123229f179bd" [ 828.007417] env[65788]: _type = "HttpNfcLease" [ 828.007417] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 828.007417] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 828.007417] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b9f79-ea5c-e56b-9bcb-123229f179bd" [ 828.007417] env[65788]: _type = "HttpNfcLease" [ 828.007417] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 828.008416] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ddf7f6-4184-452c-8872-135aeba15008 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.024797] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d406c4-da23-9a62-db07-6a1350292132/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 828.025055] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d406c4-da23-9a62-db07-6a1350292132/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 828.091274] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.091420] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.091589] env[65788]: DEBUG nova.network.neutron [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 828.099213] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.099474] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.117010] env[65788]: DEBUG oslo_concurrency.lockutils [None req-85b2c6df-9eff-493c-bfc6-99a0ebde37ce tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.510s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.144969] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-462705ca-c135-4b18-bb0b-8ac5b919f0cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.249312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.458337] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.458753] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.507103] env[65788]: DEBUG nova.network.neutron [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 828.563738] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.563738] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.599454] env[65788]: WARNING neutronclient.v2_0.client [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.600271] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.600771] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.608490] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 828.738822] env[65788]: WARNING neutronclient.v2_0.client [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 828.743128] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.743128] env[65788]: WARNING openstack [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.758743] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e26f21-a994-4e3f-9c1e-703d59abd15e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.768489] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c56ecf-1fc4-4911-a1a2-c29da527ae50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.806987] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515241d1-b944-43d8-bde4-6ea961c4be8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.816344] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466abb2c-eb01-452f-bd0a-d81e28cb1610 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.837738] env[65788]: DEBUG nova.compute.provider_tree [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.883987] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 828.885209] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 828.949824] env[65788]: DEBUG nova.network.neutron [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Updating instance_info_cache with network_info: [{"id": "ef83244d-23de-4589-bab0-ea354e648ac2", "address": "fa:16:3e:1c:c4:e8", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef83244d-23", "ovs_interfaceid": "ef83244d-23de-4589-bab0-ea354e648ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 829.039362] env[65788]: WARNING neutronclient.v2_0.client [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.040154] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.040512] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.153829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.167676] env[65788]: DEBUG nova.network.neutron [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance_info_cache with network_info: [{"id": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "address": "fa:16:3e:7c:87:d4", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e5e0e29-c7", "ovs_interfaceid": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 829.186329] env[65788]: DEBUG nova.compute.manager [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Received event network-changed-31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 829.186585] env[65788]: DEBUG nova.compute.manager [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Refreshing instance network info cache due to event network-changed-31daf1d5-0a84-4d17-a694-a7ab16e9b81b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 829.186949] env[65788]: DEBUG oslo_concurrency.lockutils [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Acquiring lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.187214] env[65788]: DEBUG oslo_concurrency.lockutils [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Acquired lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.187699] env[65788]: DEBUG nova.network.neutron [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Refreshing network info cache for port 31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 829.341406] env[65788]: DEBUG nova.scheduler.client.report [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 829.453922] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "refresh_cache-37471e59-1809-4df3-8c40-20921d04d18e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.454312] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Instance network_info: |[{"id": "ef83244d-23de-4589-bab0-ea354e648ac2", "address": "fa:16:3e:1c:c4:e8", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef83244d-23", "ovs_interfaceid": "ef83244d-23de-4589-bab0-ea354e648ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 829.454895] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:c4:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef83244d-23de-4589-bab0-ea354e648ac2', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.464063] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 829.464547] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.464938] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0adea873-9a62-4355-a703-3b29b947db67 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.491124] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.491124] env[65788]: value = "task-4662409" [ 829.491124] env[65788]: _type = "Task" [ 829.491124] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.501177] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662409, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.625872] env[65788]: DEBUG nova.compute.manager [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Received event network-changed-ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 829.626408] env[65788]: DEBUG nova.compute.manager [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Refreshing instance network info cache due to event network-changed-ef83244d-23de-4589-bab0-ea354e648ac2. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 829.626782] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Acquiring lock "refresh_cache-37471e59-1809-4df3-8c40-20921d04d18e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.627156] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Acquired lock "refresh_cache-37471e59-1809-4df3-8c40-20921d04d18e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.627371] env[65788]: DEBUG nova.network.neutron [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Refreshing network info cache for port ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 829.671242] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.692114] env[65788]: WARNING neutronclient.v2_0.client [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.692989] env[65788]: WARNING openstack [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.693857] env[65788]: WARNING openstack [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.757845] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.758254] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.758539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.758748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.758994] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.761520] env[65788]: INFO nova.compute.manager [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Terminating instance [ 829.821441] env[65788]: WARNING openstack [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.822059] env[65788]: WARNING openstack [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.846995] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.323s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.849812] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 46.362s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.850124] env[65788]: DEBUG nova.objects.instance [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 829.868941] env[65788]: INFO nova.scheduler.client.report [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Deleted allocations for instance 670f59c3-10b0-431f-a8b0-ef6f82a938be [ 829.893980] env[65788]: WARNING neutronclient.v2_0.client [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 829.894823] env[65788]: WARNING openstack [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.895271] env[65788]: WARNING openstack [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.007165] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662409, 'name': CreateVM_Task, 'duration_secs': 0.458566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.007165] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.007594] env[65788]: WARNING neutronclient.v2_0.client [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.008135] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.008647] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.009612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 830.010161] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21f3a037-f367-48c2-bd9a-ace86d96e9f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.017256] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 830.017256] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204d300-82ef-8833-e82b-49729bbba1d9" [ 830.017256] env[65788]: _type = "Task" [ 830.017256] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.024366] env[65788]: DEBUG nova.network.neutron [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updated VIF entry in instance network info cache for port 31daf1d5-0a84-4d17-a694-a7ab16e9b81b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 830.024366] env[65788]: DEBUG nova.network.neutron [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updating instance_info_cache with network_info: [{"id": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "address": "fa:16:3e:91:8c:07", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.247", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31daf1d5-0a", "ovs_interfaceid": "31daf1d5-0a84-4d17-a694-a7ab16e9b81b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.036228] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204d300-82ef-8833-e82b-49729bbba1d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.132180] env[65788]: WARNING neutronclient.v2_0.client [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.132971] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.133392] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.256705] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.257240] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.265675] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "refresh_cache-93e1ade8-4c57-4db9-9da3-388e3eea91bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.267069] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquired lock "refresh_cache-93e1ade8-4c57-4db9-9da3-388e3eea91bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.267069] env[65788]: DEBUG nova.network.neutron [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 830.328915] env[65788]: WARNING neutronclient.v2_0.client [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.329689] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.330101] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.383067] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9ff8c5dd-7021-4d48-8005-31fc3fc65d4d tempest-ServersTestJSON-1142340636 tempest-ServersTestJSON-1142340636-project-member] Lock "670f59c3-10b0-431f-a8b0-ef6f82a938be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 53.686s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.425149] env[65788]: DEBUG nova.network.neutron [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Updated VIF entry in instance network info cache for port ef83244d-23de-4589-bab0-ea354e648ac2. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 830.425629] env[65788]: DEBUG nova.network.neutron [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Updating instance_info_cache with network_info: [{"id": "ef83244d-23de-4589-bab0-ea354e648ac2", "address": "fa:16:3e:1c:c4:e8", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef83244d-23", "ovs_interfaceid": "ef83244d-23de-4589-bab0-ea354e648ac2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.527895] env[65788]: DEBUG oslo_concurrency.lockutils [req-db23f180-9685-48dc-8b39-e20c0397c6f6 req-09d7bb96-5624-40a8-b712-31769b494a6e service nova] Releasing lock "refresh_cache-f160e4a5-1e91-495e-800e-116ec435d8e1" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.536270] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5204d300-82ef-8833-e82b-49729bbba1d9, 'name': SearchDatastore_Task, 'duration_secs': 0.017757} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.536598] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.536836] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.537097] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.537243] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.537425] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.537712] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-caea1acf-1a2b-4026-b49a-1a659443324c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.548431] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.548647] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 830.549584] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff23c500-4af9-4ed4-b5c1-bd104df1289e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.557383] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 830.557383] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521a2773-35ba-a291-d6f4-2b24dcebf7a7" [ 830.557383] env[65788]: _type = "Task" [ 830.557383] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.568171] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521a2773-35ba-a291-d6f4-2b24dcebf7a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.770462] env[65788]: WARNING neutronclient.v2_0.client [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 830.771164] env[65788]: WARNING openstack [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.771532] env[65788]: WARNING openstack [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.804361] env[65788]: DEBUG nova.network.neutron [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 830.856560] env[65788]: DEBUG nova.network.neutron [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 830.865944] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b7de556e-f673-4652-9e57-63e200db1da6 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.867251] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.187s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.867375] env[65788]: DEBUG nova.objects.instance [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lazy-loading 'resources' on Instance uuid f3c53f67-85bf-4c18-9313-75eb90862f78 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.928695] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Releasing lock "refresh_cache-37471e59-1809-4df3-8c40-20921d04d18e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 830.929253] env[65788]: DEBUG nova.compute.manager [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Received event network-changed-773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 830.929253] env[65788]: DEBUG nova.compute.manager [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Refreshing instance network info cache due to event network-changed-773392a6-7451-4cad-95dd-aba2407f0497. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 830.929388] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Acquiring lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.929772] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Acquired lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.929772] env[65788]: DEBUG nova.network.neutron [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Refreshing network info cache for port 773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 831.072028] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521a2773-35ba-a291-d6f4-2b24dcebf7a7, 'name': SearchDatastore_Task, 'duration_secs': 0.035931} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.075982] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dd06ce5-7099-4e40-bd0a-e76ed1d8d128 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.083826] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 831.083826] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529f14b1-7c9a-e8a2-98a4-77ad07446e82" [ 831.083826] env[65788]: _type = "Task" [ 831.083826] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.094141] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529f14b1-7c9a-e8a2-98a4-77ad07446e82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.189933] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240e4ef7-6736-4c99-b8b2-277446208df3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.209716] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 831.362162] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Releasing lock "refresh_cache-93e1ade8-4c57-4db9-9da3-388e3eea91bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.362162] env[65788]: DEBUG nova.compute.manager [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 831.362162] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.362162] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d768eb7-f45c-4bdc-88f3-13d9b4610e37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.371507] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.372024] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40634a69-a54e-40a5-a70d-5f9eaecb60e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.384028] env[65788]: DEBUG oslo_vmware.api [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 831.384028] env[65788]: value = "task-4662410" [ 831.384028] env[65788]: _type = "Task" [ 831.384028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.393080] env[65788]: DEBUG oslo_vmware.api [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662410, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.434270] env[65788]: WARNING neutronclient.v2_0.client [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.434969] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.435334] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.603950] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529f14b1-7c9a-e8a2-98a4-77ad07446e82, 'name': SearchDatastore_Task, 'duration_secs': 0.02686} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.604292] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.604535] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 37471e59-1809-4df3-8c40-20921d04d18e/37471e59-1809-4df3-8c40-20921d04d18e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 831.604817] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02b7a594-984a-4273-9d0d-e04ea7aff6b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.615685] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 831.615685] env[65788]: value = "task-4662411" [ 831.615685] env[65788]: _type = "Task" [ 831.615685] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.622392] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.622749] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.637413] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.716079] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.716400] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-722cd881-8b57-47f4-879e-62113303afb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.725498] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 831.725498] env[65788]: value = "task-4662412" [ 831.725498] env[65788]: _type = "Task" [ 831.725498] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.735904] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662412, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.803236] env[65788]: WARNING neutronclient.v2_0.client [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.803644] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.804041] env[65788]: WARNING openstack [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.893491] env[65788]: DEBUG oslo_vmware.api [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662410, 'name': PowerOffVM_Task, 'duration_secs': 0.157602} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.897194] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 831.897986] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 831.897986] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8336454-f325-42e5-815b-345aa7a29352 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.921013] env[65788]: DEBUG nova.network.neutron [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Updated VIF entry in instance network info cache for port 773392a6-7451-4cad-95dd-aba2407f0497. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 831.921622] env[65788]: DEBUG nova.network.neutron [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Updating instance_info_cache with network_info: [{"id": "773392a6-7451-4cad-95dd-aba2407f0497", "address": "fa:16:3e:04:ba:6b", "network": {"id": "6cc1fda5-0237-4c52-96c9-9f5da8aa6cf6", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-21558233-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6dcc0ac928ee42b9a926c98464c4ec95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4a8f5363-be3a-4f92-9ccf-33bb0c8113b3", "external-id": "nsx-vlan-transportzone-944", "segmentation_id": 944, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap773392a6-74", "ovs_interfaceid": "773392a6-7451-4cad-95dd-aba2407f0497", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 831.928336] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 831.928624] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 831.928842] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Deleting the datastore file [datastore2] 93e1ade8-4c57-4db9-9da3-388e3eea91bd {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.929176] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81d95c46-e8dd-47e4-a3dd-ffbb2788ce86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.940814] env[65788]: DEBUG oslo_vmware.api [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for the task: (returnval){ [ 831.940814] env[65788]: value = "task-4662414" [ 831.940814] env[65788]: _type = "Task" [ 831.940814] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.950836] env[65788]: DEBUG oslo_vmware.api [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662414, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.968484] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8f23bb-6d60-4f02-8b39-5cd4dc0a24ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.980854] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a3bcdd-745b-40db-ac7a-68151a6a997c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.030074] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954d94a8-2123-446f-853e-943653d6017f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.039534] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319fb9e0-ede9-406c-ac31-418ba13c5ad4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.055970] env[65788]: DEBUG nova.compute.provider_tree [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.128923] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662411, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.236258] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662412, 'name': PowerOffVM_Task, 'duration_secs': 0.316218} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.236556] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.236742] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 832.425789] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6ebe520-b4e8-4d66-b7d3-5456b6b18a7e req-6286ceb0-f222-4b98-b9ca-64e719f0c630 service nova] Releasing lock "refresh_cache-bbf6a242-9873-44b9-8938-2f8df11a1018" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.450732] env[65788]: DEBUG oslo_vmware.api [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Task: {'id': task-4662414, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.466722} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.450862] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.451596] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 832.451596] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.451596] env[65788]: INFO nova.compute.manager [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Took 1.09 seconds to destroy the instance on the hypervisor. [ 832.451596] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 832.451855] env[65788]: DEBUG nova.compute.manager [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 832.451988] env[65788]: DEBUG nova.network.neutron [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 832.452238] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 832.452779] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 832.453050] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.488109] env[65788]: DEBUG nova.network.neutron [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 832.488239] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 832.563995] env[65788]: DEBUG nova.scheduler.client.report [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.629408] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662411, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.773489} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.629711] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 37471e59-1809-4df3-8c40-20921d04d18e/37471e59-1809-4df3-8c40-20921d04d18e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.629934] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.630221] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed06ff0c-eeeb-4380-9601-3007ac10687a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.639195] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 832.639195] env[65788]: value = "task-4662415" [ 832.639195] env[65788]: _type = "Task" [ 832.639195] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.648973] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662415, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.743967] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 832.744323] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 832.744484] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 832.744673] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 832.744821] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 832.744980] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 832.745216] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.745378] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 832.745543] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 832.745718] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 832.745901] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 832.751338] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cb356ce-a875-4ce7-a232-994a3eb5b2e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.768693] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 832.768693] env[65788]: value = "task-4662416" [ 832.768693] env[65788]: _type = "Task" [ 832.768693] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.779339] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662416, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.993724] env[65788]: DEBUG nova.network.neutron [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 833.072840] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.206s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.076025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.938s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.076567] env[65788]: DEBUG nova.objects.instance [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lazy-loading 'resources' on Instance uuid 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.106152] env[65788]: INFO nova.scheduler.client.report [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted allocations for instance f3c53f67-85bf-4c18-9313-75eb90862f78 [ 833.158389] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662415, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.117534} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.158833] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.160974] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc3d802-ee74-48ef-9c1a-3209e465b978 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.190936] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 37471e59-1809-4df3-8c40-20921d04d18e/37471e59-1809-4df3-8c40-20921d04d18e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.192282] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-971a9dd9-490f-49db-8acd-d2eedc98408f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.213666] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 833.213666] env[65788]: value = "task-4662417" [ 833.213666] env[65788]: _type = "Task" [ 833.213666] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.223844] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.280137] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662416, 'name': ReconfigVM_Task, 'duration_secs': 0.263947} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.280470] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 833.497924] env[65788]: INFO nova.compute.manager [-] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Took 1.05 seconds to deallocate network for instance. [ 833.617681] env[65788]: DEBUG oslo_concurrency.lockutils [None req-23b76d21-fb6b-4b3b-856a-3d796dd57ec5 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "f3c53f67-85bf-4c18-9313-75eb90862f78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.494s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.732466] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662417, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.790310] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 833.790714] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 833.790714] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 833.790938] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 833.793596] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 833.793596] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 833.793596] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.793596] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 833.793596] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 833.793875] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 833.793875] env[65788]: DEBUG nova.virt.hardware [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 833.798797] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Reconfiguring VM instance instance-0000002f to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 833.799283] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89dbb9b2-9aae-4ed5-b9d5-f3c1141da5f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.826141] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 833.826141] env[65788]: value = "task-4662418" [ 833.826141] env[65788]: _type = "Task" [ 833.826141] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.840258] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662418, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.917853] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "86ac6437-511a-4d72-aff0-0325e2d633f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.918521] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.918521] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "86ac6437-511a-4d72-aff0-0325e2d633f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.918521] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.918808] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.921454] env[65788]: INFO nova.compute.manager [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Terminating instance [ 834.009161] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.097741] env[65788]: INFO nova.compute.manager [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Rebuilding instance [ 834.148367] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f19e7e-ecb5-41ef-a191-44bf79e5cff1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.159995] env[65788]: DEBUG nova.compute.manager [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 834.161182] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1758414b-0fec-4fe8-baa8-4cab3771126b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.165258] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c3498b-0ecb-43aa-b6e3-805c19be69cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.202012] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c1b7cc-55b4-43cc-a0ad-79693a7a36d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.211349] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36286fc7-3cb4-4994-9949-7a4400baae26 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.229397] env[65788]: DEBUG nova.compute.provider_tree [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.233583] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662417, 'name': ReconfigVM_Task, 'duration_secs': 0.64247} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.234080] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 37471e59-1809-4df3-8c40-20921d04d18e/37471e59-1809-4df3-8c40-20921d04d18e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.234708] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fe51b41-09ce-4342-a912-97757b1d830d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.242952] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 834.242952] env[65788]: value = "task-4662419" [ 834.242952] env[65788]: _type = "Task" [ 834.242952] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.254651] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662419, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.339227] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662418, 'name': ReconfigVM_Task, 'duration_secs': 0.261944} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.339227] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Reconfigured VM instance instance-0000002f to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 834.340082] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e717d3e-1a01-4f68-9e12-5fd7ea7d9187 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.365480] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] f75000e6-f3bf-4951-a541-3aa185339054/f75000e6-f3bf-4951-a541-3aa185339054.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.365845] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f0fff8f-5881-4ed2-b5da-10732c92efe8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.386992] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 834.386992] env[65788]: value = "task-4662420" [ 834.386992] env[65788]: _type = "Task" [ 834.386992] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.398511] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662420, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.430265] env[65788]: DEBUG nova.compute.manager [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 834.430265] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.430831] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dda23a6-d25a-4d0d-812d-7147caf538a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.440842] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.441099] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce8054d6-a9f4-4efc-af33-3ea53ed70e5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.449416] env[65788]: DEBUG oslo_vmware.api [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 834.449416] env[65788]: value = "task-4662421" [ 834.449416] env[65788]: _type = "Task" [ 834.449416] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.459568] env[65788]: DEBUG oslo_vmware.api [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662421, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.736990] env[65788]: DEBUG nova.scheduler.client.report [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 834.756438] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662419, 'name': Rename_Task, 'duration_secs': 0.281358} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.757604] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 834.757950] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fc1e76a-ad7d-417c-8e82-7b3d37176b4d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.769039] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 834.769039] env[65788]: value = "task-4662422" [ 834.769039] env[65788]: _type = "Task" [ 834.769039] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.781470] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.901925] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662420, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.963831] env[65788]: DEBUG oslo_vmware.api [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662421, 'name': PowerOffVM_Task, 'duration_secs': 0.289796} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.964626] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.964626] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.964937] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23b58ee4-4fe7-4d50-810a-bfd55a2eb1e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.057081] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.057289] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.057514] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleting the datastore file [datastore2] 86ac6437-511a-4d72-aff0-0325e2d633f6 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.057826] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f18b5c0-ac0c-4938-a418-d9421399255d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.067103] env[65788]: DEBUG oslo_vmware.api [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 835.067103] env[65788]: value = "task-4662424" [ 835.067103] env[65788]: _type = "Task" [ 835.067103] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.076521] env[65788]: DEBUG oslo_vmware.api [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.206623] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.206958] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f567669-3810-47c4-a310-0653c8ef2df5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.217640] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 835.217640] env[65788]: value = "task-4662425" [ 835.217640] env[65788]: _type = "Task" [ 835.217640] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.230903] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.242200] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.166s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.245406] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 43.714s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.245675] env[65788]: DEBUG nova.objects.instance [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lazy-loading 'resources' on Instance uuid 5c7a1693-62f2-454e-9406-0b4a132ebf25 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.265041] env[65788]: INFO nova.scheduler.client.report [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Deleted allocations for instance 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1 [ 835.283123] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662422, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.333176] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.333176] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.402415] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662420, 'name': ReconfigVM_Task, 'duration_secs': 0.599629} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.402898] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Reconfigured VM instance instance-0000002f to attach disk [datastore1] f75000e6-f3bf-4951-a541-3aa185339054/f75000e6-f3bf-4951-a541-3aa185339054.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.403294] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 835.579081] env[65788]: DEBUG oslo_vmware.api [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280879} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.579551] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.580273] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 835.580273] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 835.580273] env[65788]: INFO nova.compute.manager [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Took 1.15 seconds to destroy the instance on the hypervisor. [ 835.582034] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 835.582300] env[65788]: DEBUG nova.compute.manager [-] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 835.582428] env[65788]: DEBUG nova.network.neutron [-] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 835.582673] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.583338] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 835.583602] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 835.659423] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 835.668247] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.668476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.731960] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662425, 'name': PowerOffVM_Task, 'duration_secs': 0.285159} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.731960] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.733481] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.734834] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9c7a931-b221-4a32-8f97-ab12de6d5bea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.744669] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 835.744669] env[65788]: value = "task-4662426" [ 835.744669] env[65788]: _type = "Task" [ 835.744669] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.757468] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 835.758283] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 835.758283] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910212', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'name': 'volume-0ec3463c-cdd6-4521-b484-dd20a13916e1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '263cd2fb-7eeb-4395-b1e9-b829940ebd35', 'attached_at': '', 'detached_at': '', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'serial': '0ec3463c-cdd6-4521-b484-dd20a13916e1'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 835.758769] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f5ae4b-0015-42a3-9d46-96433b6470a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.788131] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314ef008-e192-457a-88c9-2c8337d497bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.791052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e6b12ed-7053-4b40-82bd-abe4f1d7b0cd tempest-FloatingIPsAssociationTestJSON-1491106384 tempest-FloatingIPsAssociationTestJSON-1491106384-project-member] Lock "12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.192s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.801207] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16500726-0564-4dd4-a34d-8ba92bea0de4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.805956] env[65788]: DEBUG oslo_vmware.api [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662422, 'name': PowerOnVM_Task, 'duration_secs': 0.714443} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.806953] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 835.807190] env[65788]: INFO nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Took 8.97 seconds to spawn the instance on the hypervisor. [ 835.807373] env[65788]: DEBUG nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 835.808653] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddce659-3f62-4d9f-8613-a9a99a3fcc72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.832404] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73882244-ed92-42d6-8258-e55f57bdbad7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.837013] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 835.861230] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] The volume has not been displaced from its original location: [datastore1] volume-0ec3463c-cdd6-4521-b484-dd20a13916e1/volume-0ec3463c-cdd6-4521-b484-dd20a13916e1.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 835.867542] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Reconfiguring VM instance instance-0000002d to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 835.872918] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a47019e7-4870-4cfb-970b-ceff502aca81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.898096] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 835.898096] env[65788]: value = "task-4662427" [ 835.898096] env[65788]: _type = "Task" [ 835.898096] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.914509] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662427, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.915373] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f1c8d9-48a5-4a6d-90fd-3b8b13b3d906 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.946362] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5363da6e-a094-45ed-8e3e-419430e624c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.966141] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 836.171565] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 836.174039] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.174184] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.226810] env[65788]: DEBUG nova.compute.manager [req-2e0c8c8b-972d-4296-8cbb-8736bbd931a2 req-a0f952b2-fa30-4d52-ae24-178293b1a196 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Received event network-vif-deleted-83cd356c-b01d-42de-8d7b-a6b5cde9d80d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 836.227027] env[65788]: INFO nova.compute.manager [req-2e0c8c8b-972d-4296-8cbb-8736bbd931a2 req-a0f952b2-fa30-4d52-ae24-178293b1a196 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Neutron deleted interface 83cd356c-b01d-42de-8d7b-a6b5cde9d80d; detaching it from the instance and deleting it from the info cache [ 836.227214] env[65788]: DEBUG nova.network.neutron [req-2e0c8c8b-972d-4296-8cbb-8736bbd931a2 req-a0f952b2-fa30-4d52-ae24-178293b1a196 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.360903] env[65788]: INFO nova.compute.manager [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Took 60.61 seconds to build instance. [ 836.373457] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.406785] env[65788]: DEBUG nova.network.neutron [-] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.415399] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662427, 'name': ReconfigVM_Task, 'duration_secs': 0.246061} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.418419] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Reconfigured VM instance instance-0000002d to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 836.424883] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fbc6778-d9ed-4e14-9593-600a6fa7509a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.444864] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 836.444864] env[65788]: value = "task-4662428" [ 836.444864] env[65788]: _type = "Task" [ 836.444864] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.457069] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662428, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.477780] env[65788]: WARNING neutronclient.v2_0.client [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 836.537851] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d86f30a-f6d2-47bb-932a-99c48b70f74e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.547208] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d3bcbf-6880-4ac9-aff7-71a7544c54a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.581121] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e6037d-1af5-4983-b79e-2e9d6e9c2bf2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.590587] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b72137-ada9-4a3f-b098-806d37e35ae0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.605317] env[65788]: DEBUG nova.compute.provider_tree [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.678802] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 836.701464] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 836.730349] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c89285f0-6de8-49e5-b73e-7d81a6ae3754 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.742716] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e91a75-d028-46af-af36-f7a4dac6c03c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.789019] env[65788]: DEBUG nova.compute.manager [req-2e0c8c8b-972d-4296-8cbb-8736bbd931a2 req-a0f952b2-fa30-4d52-ae24-178293b1a196 service nova] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Detach interface failed, port_id=83cd356c-b01d-42de-8d7b-a6b5cde9d80d, reason: Instance 86ac6437-511a-4d72-aff0-0325e2d633f6 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 836.811418] env[65788]: DEBUG nova.network.neutron [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Port 9e5e0e29-c770-41e3-9766-2be8e6f3dc96 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 836.863232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb5d5c00-bf03-45c3-9646-f105b458f069 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "37471e59-1809-4df3-8c40-20921d04d18e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.559s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.919352] env[65788]: INFO nova.compute.manager [-] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Took 1.34 seconds to deallocate network for instance. [ 836.957458] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662428, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.110917] env[65788]: DEBUG nova.scheduler.client.report [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 837.222310] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.426471] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.457421] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662428, 'name': ReconfigVM_Task, 'duration_secs': 0.782308} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.457796] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910212', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'name': 'volume-0ec3463c-cdd6-4521-b484-dd20a13916e1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '263cd2fb-7eeb-4395-b1e9-b829940ebd35', 'attached_at': '', 'detached_at': '', 'volume_id': '0ec3463c-cdd6-4521-b484-dd20a13916e1', 'serial': '0ec3463c-cdd6-4521-b484-dd20a13916e1'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 837.458213] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 837.459066] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539c1145-0af2-4ea1-927b-d4e6d10f583a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.469995] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 837.470466] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-067f2523-fa02-47a2-a6bf-9ab25a787661 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.558804] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 837.559049] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 837.559238] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Deleting the datastore file [datastore1] 263cd2fb-7eeb-4395-b1e9-b829940ebd35 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.559530] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0afed235-3c90-4692-a1a0-ea4eaec59149 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.572051] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for the task: (returnval){ [ 837.572051] env[65788]: value = "task-4662430" [ 837.572051] env[65788]: _type = "Task" [ 837.572051] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.586043] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662430, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.617780] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.373s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.621325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 45.742s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.655167] env[65788]: INFO nova.scheduler.client.report [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Deleted allocations for instance 5c7a1693-62f2-454e-9406-0b4a132ebf25 [ 837.836920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "f75000e6-f3bf-4951-a541-3aa185339054-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.837233] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.837415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.083837] env[65788]: DEBUG oslo_vmware.api [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Task: {'id': task-4662430, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132293} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.084236] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.084431] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 838.084626] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 838.164390] env[65788]: DEBUG oslo_concurrency.lockutils [None req-245817d2-8cf2-4f03-b598-7419b5b4fbf6 tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "5c7a1693-62f2-454e-9406-0b4a132ebf25" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.605s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.167987] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 838.168436] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-349c6b64-d768-4178-9fcb-02d04737fec3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.186323] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5b68a0-1e98-484a-87c2-b9917afe329a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.238266] env[65788]: ERROR nova.compute.manager [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Failed to detach volume 0ec3463c-cdd6-4521-b484-dd20a13916e1 from /dev/sda: nova.exception.InstanceNotFound: Instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 could not be found. [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Traceback (most recent call last): [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 4224, in _do_rebuild_instance [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self.driver.rebuild(**kwargs) [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] raise NotImplementedError() [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] NotImplementedError [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] During handling of the above exception, another exception occurred: [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Traceback (most recent call last): [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3647, in _detach_root_volume [ 838.238266] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self.driver.detach_volume(context, old_connection_info, [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] return self._volumeops.detach_volume(connection_info, instance) [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._detach_volume_vmdk(connection_info, instance) [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] stable_ref.fetch_moref(session) [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] raise exception.InstanceNotFound(instance_id=self._uuid) [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] nova.exception.InstanceNotFound: Instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 could not be found. [ 838.238790] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.519445] env[65788]: DEBUG nova.compute.utils [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Build of instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 aborted: Failed to rebuild volume backed instance. {{(pid=65788) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 838.527168] env[65788]: ERROR nova.compute.manager [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 aborted: Failed to rebuild volume backed instance. [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Traceback (most recent call last): [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 4224, in _do_rebuild_instance [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self.driver.rebuild(**kwargs) [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/driver.py", line 533, in rebuild [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] raise NotImplementedError() [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] NotImplementedError [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] During handling of the above exception, another exception occurred: [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Traceback (most recent call last): [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3682, in _rebuild_volume_backed_instance [ 838.527168] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._detach_root_volume(context, instance, root_bdm) [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3661, in _detach_root_volume [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] with excutils.save_and_reraise_exception(): [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self.force_reraise() [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] raise self.value [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3647, in _detach_root_volume [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self.driver.detach_volume(context, old_connection_info, [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] return self._volumeops.detach_volume(connection_info, instance) [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 838.527806] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._detach_volume_vmdk(connection_info, instance) [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] stable_ref.fetch_moref(session) [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] raise exception.InstanceNotFound(instance_id=self._uuid) [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] nova.exception.InstanceNotFound: Instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 could not be found. [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] During handling of the above exception, another exception occurred: [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Traceback (most recent call last): [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 11553, in _error_out_instance_on_exception [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] yield [ 838.528222] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3950, in rebuild_instance [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._do_rebuild_instance_with_claim( [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 4036, in _do_rebuild_instance_with_claim [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._do_rebuild_instance( [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 4228, in _do_rebuild_instance [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._rebuild_default_impl(**kwargs) [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3805, in _rebuild_default_impl [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] self._rebuild_volume_backed_instance( [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] File "/opt/stack/nova/nova/compute/manager.py", line 3697, in _rebuild_volume_backed_instance [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] raise exception.BuildAbortException( [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] nova.exception.BuildAbortException: Build of instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 aborted: Failed to rebuild volume backed instance. [ 838.528653] env[65788]: ERROR nova.compute.manager [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] [ 838.665213] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d406c4-da23-9a62-db07-6a1350292132/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 838.666527] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d59710a-9aa8-4f69-98e4-1e89a9897b06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.673870] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d406c4-da23-9a62-db07-6a1350292132/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 838.674120] env[65788]: ERROR oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d406c4-da23-9a62-db07-6a1350292132/disk-0.vmdk due to incomplete transfer. [ 838.674369] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-34b7a646-e5e0-4dbb-ad49-8a981e0f818b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.687345] env[65788]: DEBUG oslo_vmware.rw_handles [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d406c4-da23-9a62-db07-6a1350292132/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 838.687753] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Uploaded image aec2bed9-f33a-4c13-860a-a321f69ec40b to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 838.689880] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 838.692696] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c9f04197-80c9-4984-9840-ffc4834b061f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.700884] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 838.700884] env[65788]: value = "task-4662431" [ 838.700884] env[65788]: _type = "Task" [ 838.700884] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.714118] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662431, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.715285] env[65788]: DEBUG nova.compute.manager [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 838.716163] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d825077a-45e5-46e6-8d7e-800a9bb0dcdb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.764638] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b494ad-0465-45e3-8c6a-382d65ce61d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.775407] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc2f0a6-4d0e-45c8-9fba-db2b76ebd448 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.808506] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c9e192-feb2-4972-9bf2-a7505af7ab87 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.817506] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27479093-2e5d-4f17-9d14-f4b5fe504f5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.833267] env[65788]: DEBUG nova.compute.provider_tree [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.845866] env[65788]: WARNING neutronclient.v2_0.client [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 838.883360] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "1ef08e8f-744d-4939-9433-a4301cd31e77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.883581] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.883786] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "1ef08e8f-744d-4939-9433-a4301cd31e77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.883967] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.884243] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.886404] env[65788]: INFO nova.compute.manager [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Terminating instance [ 838.892111] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.892111] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.892111] env[65788]: DEBUG nova.network.neutron [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 839.210737] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662431, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.227298] env[65788]: INFO nova.compute.manager [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] instance snapshotting [ 839.234045] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c319058-71f2-4204-bb8e-ac2b9e150bf8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.259708] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ccb4ec-c91b-448e-855a-cb691918600f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.336711] env[65788]: DEBUG nova.scheduler.client.report [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 839.390744] env[65788]: DEBUG nova.compute.manager [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 839.391095] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.392062] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cab884-60af-4aa3-b4d9-e082ea5471a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.395555] env[65788]: WARNING neutronclient.v2_0.client [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 839.397121] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.397121] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.410978] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.412970] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30f66c66-76eb-4f5d-9b14-32d4c616377f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.422288] env[65788]: DEBUG oslo_vmware.api [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 839.422288] env[65788]: value = "task-4662432" [ 839.422288] env[65788]: _type = "Task" [ 839.422288] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.432250] env[65788]: DEBUG oslo_vmware.api [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.538771] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.539358] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.693328] env[65788]: WARNING neutronclient.v2_0.client [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 839.693328] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 839.693328] env[65788]: WARNING openstack [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 839.713982] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662431, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.777797] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 839.778259] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-23cab08a-3934-43f3-a3e5-76231e3fcc1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.787240] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 839.787240] env[65788]: value = "task-4662433" [ 839.787240] env[65788]: _type = "Task" [ 839.787240] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.797301] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662433, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.806363] env[65788]: DEBUG nova.network.neutron [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance_info_cache with network_info: [{"id": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "address": "fa:16:3e:7c:87:d4", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e5e0e29-c7", "ovs_interfaceid": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.933631] env[65788]: DEBUG oslo_vmware.api [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662432, 'name': PowerOffVM_Task, 'duration_secs': 0.295163} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.933905] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.934090] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.934417] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb0120ea-8a92-48e9-b6ef-8b8cd7714faf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.014255] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 840.014479] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 840.014673] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Deleting the datastore file [datastore1] 1ef08e8f-744d-4939-9433-a4301cd31e77 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.015036] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-878f2f8c-9bba-4b5b-b77f-b28f7db6307e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.024377] env[65788]: DEBUG oslo_vmware.api [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for the task: (returnval){ [ 840.024377] env[65788]: value = "task-4662435" [ 840.024377] env[65788]: _type = "Task" [ 840.024377] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.034731] env[65788]: DEBUG oslo_vmware.api [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662435, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.212575] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662431, 'name': Destroy_Task, 'duration_secs': 1.052134} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.212575] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Destroyed the VM [ 840.212575] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 840.212802] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-69d760b1-03e4-422b-83a2-5115f9cfa096 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.220313] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 840.220313] env[65788]: value = "task-4662436" [ 840.220313] env[65788]: _type = "Task" [ 840.220313] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.230466] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662436, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.297550] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662433, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.309405] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 840.350161] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.729s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.350161] env[65788]: DEBUG nova.compute.manager [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=65788) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5416}} [ 840.353125] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 41.665s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 840.355029] env[65788]: INFO nova.compute.claims [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.536864] env[65788]: DEBUG oslo_vmware.api [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Task: {'id': task-4662435, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239686} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.537185] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.537390] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.537585] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.537766] env[65788]: INFO nova.compute.manager [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Took 1.15 seconds to destroy the instance on the hypervisor. [ 840.538067] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 840.538278] env[65788]: DEBUG nova.compute.manager [-] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 840.538377] env[65788]: DEBUG nova.network.neutron [-] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 840.538723] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.542605] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.542605] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.554025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.582230] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 840.732175] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662436, 'name': RemoveSnapshot_Task} progress is 26%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.802221] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662433, 'name': CreateSnapshot_Task, 'duration_secs': 0.724552} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.802221] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 840.802981] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48140fe7-fa85-4c69-a6a8-a5e1b8bec36e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.853701] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2f5b5d-b701-4a56-a520-c8ea37ed026d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.891480] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71210f78-e362-42a6-b146-b73a0d96b18e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.907340] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 840.987515] env[65788]: INFO nova.scheduler.client.report [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted allocation for migration 7fa860c2-3230-490e-9d69-7e7474adafa0 [ 841.171580] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.172044] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.172332] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.172587] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.172848] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.176575] env[65788]: INFO nova.compute.manager [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Terminating instance [ 841.223506] env[65788]: DEBUG nova.compute.manager [req-c9876a24-04c0-444e-9124-7472d673912f req-54187cd3-ee9f-4b2d-b389-5c8d19ce3c27 service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Received event network-vif-deleted-619462da-ce8b-4b5f-9459-5229925adbb6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 841.223771] env[65788]: INFO nova.compute.manager [req-c9876a24-04c0-444e-9124-7472d673912f req-54187cd3-ee9f-4b2d-b389-5c8d19ce3c27 service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Neutron deleted interface 619462da-ce8b-4b5f-9459-5229925adbb6; detaching it from the instance and deleting it from the info cache [ 841.224017] env[65788]: DEBUG nova.network.neutron [req-c9876a24-04c0-444e-9124-7472d673912f req-54187cd3-ee9f-4b2d-b389-5c8d19ce3c27 service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.238026] env[65788]: DEBUG oslo_vmware.api [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662436, 'name': RemoveSnapshot_Task, 'duration_secs': 0.773948} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.238026] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 841.238026] env[65788]: INFO nova.compute.manager [None req-256d8540-344b-40fc-8254-df71486a142e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Took 17.43 seconds to snapshot the instance on the hypervisor. [ 841.326717] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 841.330108] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3fd44072-caa1-42e1-8b13-84b45c2faf0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.343013] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 841.343013] env[65788]: value = "task-4662437" [ 841.343013] env[65788]: _type = "Task" [ 841.343013] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.353523] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662437, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.397276] env[65788]: DEBUG nova.network.neutron [-] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.417870] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.418844] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-12caf41e-d531-40cc-a2fb-e18ae582be73 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.431330] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 841.431330] env[65788]: value = "task-4662438" [ 841.431330] env[65788]: _type = "Task" [ 841.431330] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.444158] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.453504] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe972b6f-c162-4ac9-929e-154f8a96ea58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.461687] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.461961] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.470316] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9cd039-8e7a-482d-95ad-4c7817dd955f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.510346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e9dfaf53-d010-42a5-9d27-c0c840ae1a2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 52.734s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.512562] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6690cb5e-6ae9-4648-bf34-746f30f02672 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.523294] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8403f6-648f-45e6-a5ea-851605ba342b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.540755] env[65788]: DEBUG nova.compute.provider_tree [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.681853] env[65788]: DEBUG nova.compute.manager [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 841.681853] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b12ffd59-09a1-4679-9d27-3702f3135441 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.692383] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12084d77-14bb-4e0b-ad1a-08affe0b1748 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.731974] env[65788]: WARNING nova.virt.vmwareapi.driver [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 could not be found. [ 841.732328] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.733296] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8a1ee49-6f75-4a45-8109-ae799266cdf3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.735570] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10c72909-50b0-4729-b207-03f720720713 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.751924] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c74040-0547-46d1-978f-cc6a90ca3fc2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.776046] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd383ec-2240-42ee-94fb-41bf30b717b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.814857] env[65788]: DEBUG nova.compute.manager [req-c9876a24-04c0-444e-9124-7472d673912f req-54187cd3-ee9f-4b2d-b389-5c8d19ce3c27 service nova] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Detach interface failed, port_id=619462da-ce8b-4b5f-9459-5229925adbb6, reason: Instance 1ef08e8f-744d-4939-9433-a4301cd31e77 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 841.839717] env[65788]: WARNING nova.virt.vmwareapi.vmops [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 could not be found. [ 841.840412] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.840831] env[65788]: INFO nova.compute.manager [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Took 0.16 seconds to destroy the instance on the hypervisor. [ 841.842018] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 841.843053] env[65788]: DEBUG nova.compute.manager [-] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 841.843053] env[65788]: DEBUG nova.network.neutron [-] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 841.843053] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 841.844124] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 841.844124] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 841.863050] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662437, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.894570] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 841.902634] env[65788]: INFO nova.compute.manager [-] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Took 1.36 seconds to deallocate network for instance. [ 841.949305] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662438, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.965260] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 842.044828] env[65788]: DEBUG nova.scheduler.client.report [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.215769] env[65788]: DEBUG nova.objects.instance [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'flavor' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.364569] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662437, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.410884] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.445812] env[65788]: DEBUG oslo_vmware.api [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662438, 'name': PowerOnVM_Task, 'duration_secs': 0.695288} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.446334] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.446334] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c70e4b83-e099-46b7-9a60-11d5d212f4f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance 'f75000e6-f3bf-4951-a541-3aa185339054' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 842.496309] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.551024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.198s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.551930] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 842.557650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 39.626s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.558101] env[65788]: DEBUG nova.objects.instance [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 842.680829] env[65788]: DEBUG nova.compute.manager [req-c41547af-dc5b-41eb-918e-52b1810a334e req-029db0c5-fbeb-4fac-9fa2-8422c7649a1d service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Received event network-vif-deleted-82d30609-6b15-4172-aef8-f9afa1ed72e9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 842.681010] env[65788]: INFO nova.compute.manager [req-c41547af-dc5b-41eb-918e-52b1810a334e req-029db0c5-fbeb-4fac-9fa2-8422c7649a1d service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Neutron deleted interface 82d30609-6b15-4172-aef8-f9afa1ed72e9; detaching it from the instance and deleting it from the info cache [ 842.681297] env[65788]: DEBUG nova.network.neutron [req-c41547af-dc5b-41eb-918e-52b1810a334e req-029db0c5-fbeb-4fac-9fa2-8422c7649a1d service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 842.725443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.725624] env[65788]: DEBUG oslo_concurrency.lockutils [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 842.725831] env[65788]: DEBUG nova.network.neutron [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 842.726593] env[65788]: DEBUG nova.objects.instance [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'info_cache' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.862359] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662437, 'name': CloneVM_Task, 'duration_secs': 1.473237} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.862649] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Created linked-clone VM from snapshot [ 842.864236] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bed2157-5a09-4ef9-927f-45a8af212be6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.873285] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Uploading image e6f1a08c-b319-4a8d-9b85-bd96892a6aab {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 842.900053] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 842.900053] env[65788]: value = "vm-910286" [ 842.900053] env[65788]: _type = "VirtualMachine" [ 842.900053] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 842.900839] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-883a8fdb-439f-4dbf-9aaf-ea685105ab60 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.905391] env[65788]: DEBUG nova.network.neutron [-] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 842.910555] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease: (returnval){ [ 842.910555] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3a82b-cb2c-e404-2209-bebef9d4d8c6" [ 842.910555] env[65788]: _type = "HttpNfcLease" [ 842.910555] env[65788]: } obtained for exporting VM: (result){ [ 842.910555] env[65788]: value = "vm-910286" [ 842.910555] env[65788]: _type = "VirtualMachine" [ 842.910555] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 842.911024] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the lease: (returnval){ [ 842.911024] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3a82b-cb2c-e404-2209-bebef9d4d8c6" [ 842.911024] env[65788]: _type = "HttpNfcLease" [ 842.911024] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 842.925295] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 842.925295] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3a82b-cb2c-e404-2209-bebef9d4d8c6" [ 842.925295] env[65788]: _type = "HttpNfcLease" [ 842.925295] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 843.065995] env[65788]: DEBUG nova.compute.utils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 843.070763] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 843.071069] env[65788]: DEBUG nova.network.neutron [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 843.071348] env[65788]: WARNING neutronclient.v2_0.client [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.071789] env[65788]: WARNING neutronclient.v2_0.client [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.072533] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.073088] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.083019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3d251f8d-1a1b-4815-ab99-c8dd891cee92 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.524s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.084514] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.788s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.087249] env[65788]: INFO nova.compute.claims [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.144324] env[65788]: DEBUG nova.policy [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9a17b2eb1434b0e9297442bfc28d152', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e7aa18dcc41403389518d700c2c8b8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 843.189018] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-662f2920-8a9d-4dd7-a624-04e7af25d8a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.201019] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ee8acd-4ef3-43f1-82fa-061da30939f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.247123] env[65788]: DEBUG nova.objects.base [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 843.250397] env[65788]: DEBUG nova.compute.manager [req-c41547af-dc5b-41eb-918e-52b1810a334e req-029db0c5-fbeb-4fac-9fa2-8422c7649a1d service nova] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Detach interface failed, port_id=82d30609-6b15-4172-aef8-f9afa1ed72e9, reason: Instance 263cd2fb-7eeb-4395-b1e9-b829940ebd35 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 843.409489] env[65788]: INFO nova.compute.manager [-] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Took 1.57 seconds to deallocate network for instance. [ 843.421207] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.421207] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3a82b-cb2c-e404-2209-bebef9d4d8c6" [ 843.421207] env[65788]: _type = "HttpNfcLease" [ 843.421207] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 843.421522] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 843.421522] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d3a82b-cb2c-e404-2209-bebef9d4d8c6" [ 843.421522] env[65788]: _type = "HttpNfcLease" [ 843.421522] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 843.422309] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd40660-2a73-477a-a824-caaf72d7b52f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.431755] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e143ec-9f7c-8d04-6fad-7a85bfcb68c4/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 843.431755] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e143ec-9f7c-8d04-6fad-7a85bfcb68c4/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 843.493053] env[65788]: DEBUG nova.network.neutron [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Successfully created port: 2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 843.532622] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7d67bca1-8984-47bb-a979-5199fac6c3ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.549870] env[65788]: DEBUG nova.compute.manager [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 843.550955] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79ba6e6-ecea-494a-9f50-7027be830a3a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.572428] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 843.754032] env[65788]: WARNING neutronclient.v2_0.client [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 843.754032] env[65788]: WARNING openstack [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.754032] env[65788]: WARNING openstack [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 844.020251] env[65788]: INFO nova.compute.manager [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Took 0.61 seconds to detach 1 volumes for instance. [ 844.021149] env[65788]: DEBUG nova.compute.manager [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Deleting volume: 0ec3463c-cdd6-4521-b484-dd20a13916e1 {{(pid=65788) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3320}} [ 844.073304] env[65788]: INFO nova.compute.manager [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] instance snapshotting [ 844.076926] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a418bce3-3d55-4023-9164-421664a2903d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.107237] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5412c36e-ca21-4ded-8d38-898aa0dc389a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.577891] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.607567] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 844.618178] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5620527c-48c0-4959-a7bf-8921e0a6fc30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.622351] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 844.623183] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9f2b84ec-18f4-4323-b65d-81422d389dd8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.633988] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ce0964-3cba-4ab1-81c3-83e5a95c7476 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.638257] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 844.638257] env[65788]: value = "task-4662441" [ 844.638257] env[65788]: _type = "Task" [ 844.638257] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.641925] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 844.642344] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 844.642623] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 844.642915] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 844.643167] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 844.643418] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 844.643755] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.644017] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 844.644310] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 844.644567] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 844.644854] env[65788]: DEBUG nova.virt.hardware [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 844.646036] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2308c1e-1d06-46ad-b67d-fa4e39310c7a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.685729] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca51108-f1eb-4190-b532-aedb8218e2f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.694682] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662441, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.699296] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57a6d55-f095-4b6d-a99f-7e001365baba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.711872] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0af5eba-b4dc-47d3-93aa-e18a90833deb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.754454] env[65788]: DEBUG nova.compute.provider_tree [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.895700] env[65788]: WARNING openstack [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 844.896182] env[65788]: WARNING openstack [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.004222] env[65788]: WARNING neutronclient.v2_0.client [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 845.005144] env[65788]: WARNING openstack [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 845.005732] env[65788]: WARNING openstack [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 845.162043] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662441, 'name': CreateSnapshot_Task, 'duration_secs': 0.525649} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.162397] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 845.163219] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56659e68-88cf-4ce3-a6de-9a7766973a93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.215381] env[65788]: DEBUG nova.network.neutron [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Successfully updated port: 2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 845.259138] env[65788]: DEBUG nova.scheduler.client.report [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 845.298592] env[65788]: DEBUG nova.network.neutron [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [{"id": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "address": "fa:16:3e:40:57:3a", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e04a22-9f", "ovs_interfaceid": "d6e04a22-9f86-43c9-ac42-95191eaf454b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 845.525093] env[65788]: DEBUG nova.compute.manager [req-cbbd754b-3d5d-458b-af94-bcf6493b8ffb req-b21c9ef1-70d4-42b1-954f-bce8ee08aa54 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-vif-plugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 845.525652] env[65788]: DEBUG oslo_concurrency.lockutils [req-cbbd754b-3d5d-458b-af94-bcf6493b8ffb req-b21c9ef1-70d4-42b1-954f-bce8ee08aa54 service nova] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.527751] env[65788]: DEBUG oslo_concurrency.lockutils [req-cbbd754b-3d5d-458b-af94-bcf6493b8ffb req-b21c9ef1-70d4-42b1-954f-bce8ee08aa54 service nova] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.527751] env[65788]: DEBUG oslo_concurrency.lockutils [req-cbbd754b-3d5d-458b-af94-bcf6493b8ffb req-b21c9ef1-70d4-42b1-954f-bce8ee08aa54 service nova] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.527751] env[65788]: DEBUG nova.compute.manager [req-cbbd754b-3d5d-458b-af94-bcf6493b8ffb req-b21c9ef1-70d4-42b1-954f-bce8ee08aa54 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] No waiting events found dispatching network-vif-plugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 845.527751] env[65788]: WARNING nova.compute.manager [req-cbbd754b-3d5d-458b-af94-bcf6493b8ffb req-b21c9ef1-70d4-42b1-954f-bce8ee08aa54 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received unexpected event network-vif-plugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 for instance with vm_state building and task_state spawning. [ 845.683846] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 845.685176] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5c0a1076-f89d-46d1-8f9f-2fbd03af91c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.696792] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 845.696792] env[65788]: value = "task-4662442" [ 845.696792] env[65788]: _type = "Task" [ 845.696792] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.708373] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662442, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.720274] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.720502] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.721640] env[65788]: DEBUG nova.network.neutron [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 845.767742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.684s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.768355] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 845.771316] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 33.230s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.801810] env[65788]: DEBUG oslo_concurrency.lockutils [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-eceafff8-7d28-4b9b-ade6-5736d5977efa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.152262] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "f75000e6-f3bf-4951-a541-3aa185339054" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 846.152572] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.152768] env[65788]: DEBUG nova.compute.manager [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Going to confirm migration 4 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 846.212235] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662442, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.225328] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.225749] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.275161] env[65788]: DEBUG nova.compute.utils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 846.277383] env[65788]: DEBUG nova.network.neutron [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 846.279753] env[65788]: DEBUG nova.objects.instance [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lazy-loading 'migration_context' on Instance uuid f4a490a2-f2b1-4eac-8c1a-a18758583c70 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.281347] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 846.281347] env[65788]: DEBUG nova.network.neutron [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 846.281539] env[65788]: WARNING neutronclient.v2_0.client [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.282173] env[65788]: WARNING neutronclient.v2_0.client [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.282544] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.283153] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.337189] env[65788]: DEBUG nova.policy [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9310d3b6bf5430289df41fc742584c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4cffe8dfc424459a90e316ac4ac905f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 846.348997] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.349612] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.662202] env[65788]: WARNING neutronclient.v2_0.client [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.710918] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662442, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.743166] env[65788]: DEBUG nova.network.neutron [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Successfully created port: 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 846.786731] env[65788]: WARNING neutronclient.v2_0.client [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.788900] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.788900] env[65788]: WARNING openstack [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.803204] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 846.810647] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 846.811036] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a0d6634-f393-4a32-9d80-b72110cd972c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.824446] env[65788]: DEBUG oslo_vmware.api [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 846.824446] env[65788]: value = "task-4662443" [ 846.824446] env[65788]: _type = "Task" [ 846.824446] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.837912] env[65788]: DEBUG oslo_vmware.api [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662443, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.844090] env[65788]: WARNING neutronclient.v2_0.client [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 846.844624] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.845823] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.845823] env[65788]: DEBUG nova.network.neutron [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 846.845823] env[65788]: DEBUG nova.objects.instance [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'info_cache' on Instance uuid f75000e6-f3bf-4951-a541-3aa185339054 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.212228] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662442, 'name': CloneVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.326328] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698ceb9d-45ff-4199-853c-740529e3c9b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.343681] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e66d30-4e45-4f09-9f6f-8fc367a1b1fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.347364] env[65788]: DEBUG oslo_vmware.api [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662443, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.391627] env[65788]: WARNING neutronclient.v2_0.client [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 847.392718] env[65788]: WARNING openstack [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.393414] env[65788]: WARNING openstack [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.401305] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c343ba-e262-428e-ab09-fa8b53d53877 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.412355] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f634b51-b0bb-4a77-b606-4450d3839cef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.428377] env[65788]: DEBUG nova.compute.provider_tree [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.711696] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662442, 'name': CloneVM_Task, 'duration_secs': 1.541881} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.712096] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Created linked-clone VM from snapshot [ 847.713077] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda087c9-3faa-4732-bb8f-72d5534f9719 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.722120] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Uploading image b9915b20-93e4-4ad5-851f-b711fc0ab2c2 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 847.748890] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 847.748890] env[65788]: value = "vm-910288" [ 847.748890] env[65788]: _type = "VirtualMachine" [ 847.748890] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 847.750429] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9c290965-220b-4204-90a8-d4e3d547b2ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.758546] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lease: (returnval){ [ 847.758546] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52791083-380e-09aa-031d-4f9f04898937" [ 847.758546] env[65788]: _type = "HttpNfcLease" [ 847.758546] env[65788]: } obtained for exporting VM: (result){ [ 847.758546] env[65788]: value = "vm-910288" [ 847.758546] env[65788]: _type = "VirtualMachine" [ 847.758546] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 847.759443] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the lease: (returnval){ [ 847.759443] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52791083-380e-09aa-031d-4f9f04898937" [ 847.759443] env[65788]: _type = "HttpNfcLease" [ 847.759443] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 847.767156] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 847.767156] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52791083-380e-09aa-031d-4f9f04898937" [ 847.767156] env[65788]: _type = "HttpNfcLease" [ 847.767156] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 847.817257] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 847.843198] env[65788]: DEBUG oslo_vmware.api [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662443, 'name': PowerOnVM_Task, 'duration_secs': 0.592144} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.845761] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 847.846040] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 847.846456] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 847.846689] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 847.846848] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 847.847011] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 847.847306] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.847508] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 847.847689] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 847.847862] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 847.848056] env[65788]: DEBUG nova.virt.hardware [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 847.848383] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 847.848580] env[65788]: DEBUG nova.compute.manager [None req-348fb2df-b439-465b-b59f-8a2b65bc345b tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 847.849571] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a00f76a-01c6-4f9b-855c-b82511d42f43 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.853309] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa82e01-c256-4d6a-a608-f5ce82f7359f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.864069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4d378f-9a68-42bb-ba41-d03f5a4cba83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.931939] env[65788]: DEBUG nova.scheduler.client.report [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 848.035662] env[65788]: DEBUG nova.network.neutron [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.280822] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.280822] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52791083-380e-09aa-031d-4f9f04898937" [ 848.280822] env[65788]: _type = "HttpNfcLease" [ 848.280822] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 848.281191] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 848.281191] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52791083-380e-09aa-031d-4f9f04898937" [ 848.281191] env[65788]: _type = "HttpNfcLease" [ 848.281191] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 848.284561] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7be7a0-10ad-41e5-afde-d26062e25abd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.295077] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5258f653-5850-bf0e-1398-b5752eaed473/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 848.295077] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5258f653-5850-bf0e-1398-b5752eaed473/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 848.355057] env[65788]: WARNING openstack [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 848.355458] env[65788]: WARNING openstack [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 848.425268] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8531a3d8-6d5a-4d61-a16c-db233b26bc51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.485265] env[65788]: DEBUG nova.network.neutron [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Successfully updated port: 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 848.509027] env[65788]: WARNING neutronclient.v2_0.client [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 848.509705] env[65788]: WARNING openstack [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 848.510093] env[65788]: WARNING openstack [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 848.540361] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.540751] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance network_info: |[{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 848.541322] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:d1:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cd6eb89-f768-4ee6-93a2-386b83c42638', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.551364] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating folder: Project (5e7aa18dcc41403389518d700c2c8b8a). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.552975] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce0fc459-6699-401f-93a5-7821ed8670bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.568481] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Created folder: Project (5e7aa18dcc41403389518d700c2c8b8a) in parent group-v910111. [ 848.568909] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating folder: Instances. Parent ref: group-v910289. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.569264] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c30dd7b-29da-4ee5-9d16-19e1e2edf052 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.585278] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Created folder: Instances in parent group-v910289. [ 848.585278] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 848.585500] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.585821] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9da22e49-abdd-4ab7-8f10-be869a27ce40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.608789] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.608789] env[65788]: value = "task-4662447" [ 848.608789] env[65788]: _type = "Task" [ 848.608789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.618709] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662447, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.632717] env[65788]: DEBUG nova.network.neutron [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance_info_cache with network_info: [{"id": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "address": "fa:16:3e:7c:87:d4", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e5e0e29-c7", "ovs_interfaceid": "9e5e0e29-c770-41e3-9766-2be8e6f3dc96", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.768461] env[65788]: DEBUG nova.compute.manager [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-changed-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 848.768461] env[65788]: DEBUG nova.compute.manager [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Refreshing instance network info cache due to event network-changed-2cd6eb89-f768-4ee6-93a2-386b83c42638. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 848.768805] env[65788]: DEBUG oslo_concurrency.lockutils [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.768805] env[65788]: DEBUG oslo_concurrency.lockutils [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 848.769156] env[65788]: DEBUG nova.network.neutron [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Refreshing network info cache for port 2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 848.945810] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.175s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 848.956336] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 34.786s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 848.989527] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.990659] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 848.992665] env[65788]: DEBUG nova.network.neutron [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 849.023751] env[65788]: DEBUG nova.compute.manager [req-1a66adda-e1de-45f9-89b3-1139c235b656 req-5617ea2d-3ff0-4fe0-bef8-1d3ad0a66b28 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received event network-vif-plugged-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 849.025439] env[65788]: DEBUG oslo_concurrency.lockutils [req-1a66adda-e1de-45f9-89b3-1139c235b656 req-5617ea2d-3ff0-4fe0-bef8-1d3ad0a66b28 service nova] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.025705] env[65788]: DEBUG oslo_concurrency.lockutils [req-1a66adda-e1de-45f9-89b3-1139c235b656 req-5617ea2d-3ff0-4fe0-bef8-1d3ad0a66b28 service nova] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.025875] env[65788]: DEBUG oslo_concurrency.lockutils [req-1a66adda-e1de-45f9-89b3-1139c235b656 req-5617ea2d-3ff0-4fe0-bef8-1d3ad0a66b28 service nova] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.026074] env[65788]: DEBUG nova.compute.manager [req-1a66adda-e1de-45f9-89b3-1139c235b656 req-5617ea2d-3ff0-4fe0-bef8-1d3ad0a66b28 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] No waiting events found dispatching network-vif-plugged-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 849.026243] env[65788]: WARNING nova.compute.manager [req-1a66adda-e1de-45f9-89b3-1139c235b656 req-5617ea2d-3ff0-4fe0-bef8-1d3ad0a66b28 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received unexpected event network-vif-plugged-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 for instance with vm_state building and task_state spawning. [ 849.122553] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662447, 'name': CreateVM_Task, 'duration_secs': 0.421857} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.122553] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.123131] env[65788]: WARNING neutronclient.v2_0.client [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.123852] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.124447] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.124595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 849.124951] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ba238b8-38bb-4ad2-bc77-d384c825dfc2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.133786] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 849.133786] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c0748f-5335-6d40-0e94-26604080c9e1" [ 849.133786] env[65788]: _type = "Task" [ 849.133786] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.141859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-f75000e6-f3bf-4951-a541-3aa185339054" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 849.142261] env[65788]: DEBUG nova.objects.instance [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'migration_context' on Instance uuid f75000e6-f3bf-4951-a541-3aa185339054 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.152371] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c0748f-5335-6d40-0e94-26604080c9e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.275783] env[65788]: WARNING neutronclient.v2_0.client [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.277553] env[65788]: WARNING openstack [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.278228] env[65788]: WARNING openstack [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.499022] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.499022] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.540694] env[65788]: WARNING openstack [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.541281] env[65788]: WARNING openstack [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.592645] env[65788]: DEBUG nova.network.neutron [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 849.648404] env[65788]: DEBUG nova.objects.base [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 849.649387] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c0748f-5335-6d40-0e94-26604080c9e1, 'name': SearchDatastore_Task, 'duration_secs': 0.01717} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.650490] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3f6bb6-c0eb-4ad0-ad19-52b91f87c0f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.654655] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 849.655684] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 849.655684] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.655976] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.656101] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 849.661782] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-55dbc3fb-2937-44ae-9477-d702ce39ebc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.692776] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.693206] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.700397] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dec80193-e0c9-493d-bac0-a9ec8feaeaca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.703558] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 849.704663] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 849.705539] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21c4a0df-0486-441d-8ebe-1b84ed3f2b7c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.715806] env[65788]: DEBUG oslo_vmware.api [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 849.715806] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a5943c-9198-58ba-c67c-bcd789be7f21" [ 849.715806] env[65788]: _type = "Task" [ 849.715806] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.715806] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 849.715806] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed2972-9843-5319-442e-f4a61e8bb987" [ 849.715806] env[65788]: _type = "Task" [ 849.715806] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.717268] env[65788]: WARNING neutronclient.v2_0.client [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.718063] env[65788]: WARNING openstack [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.719263] env[65788]: WARNING openstack [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.743039] env[65788]: DEBUG oslo_vmware.api [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a5943c-9198-58ba-c67c-bcd789be7f21, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.749034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.749034] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed2972-9843-5319-442e-f4a61e8bb987, 'name': SearchDatastore_Task, 'duration_secs': 0.015306} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.749515] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f800438f-3b7e-4432-92b7-8157b14e7a60 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.757138] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 849.757138] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ebd4db-f51d-3598-45d0-ea3c0ce4788f" [ 849.757138] env[65788]: _type = "Task" [ 849.757138] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.769814] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ebd4db-f51d-3598-45d0-ea3c0ce4788f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.947312] env[65788]: WARNING neutronclient.v2_0.client [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 849.948312] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 849.948743] env[65788]: WARNING openstack [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 849.975132] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Applying migration context for instance f75000e6-f3bf-4951-a541-3aa185339054 as it has an incoming, in-progress migration b5064230-67b0-4b10-b1f0-9c2db8d69267. Migration status is confirming {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 849.978630] env[65788]: INFO nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating resource usage from migration b5064230-67b0-4b10-b1f0-9c2db8d69267 [ 849.994852] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.995444] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.995920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.996244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.996545] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.999528] env[65788]: INFO nova.compute.manager [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Terminating instance [ 850.014354] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.014552] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance ed55713e-8bf1-4960-8309-47498b163d02 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.015733] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 5cef20ed-d9f0-4237-ae9d-da401b953904 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.015789] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d3e2b5f6-20f1-40ac-b92e-e110ee119178 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.015985] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6215ea56-f99c-42f3-a2d5-f4d31952c8e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.016163] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 86ac6437-511a-4d72-aff0-0325e2d633f6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 850.016163] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1ef08e8f-744d-4939-9433-a4301cd31e77 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 850.016369] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9438ab56-1b4c-4778-a608-de319ab0ee43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.016416] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d78d36e6-a542-4ba8-9e29-b8a7b28c559e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.016490] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.016589] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance eceafff8-7d28-4b9b-ade6-5736d5977efa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.016794] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance a813d102-1b77-4214-8eab-2cd66e002912 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 850.017962] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 2c24bc30-d413-4714-81c2-b657cafe94bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.018139] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 47ff3955-5d9b-4d90-b8b1-276a0ce58a21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.018249] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance caceb296-5da2-4b4d-b955-5d7a238ff939 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 850.018413] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 93e1ade8-4c57-4db9-9da3-388e3eea91bd is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 850.018537] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance bbf6a242-9873-44b9-8938-2f8df11a1018 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.023875] env[65788]: DEBUG nova.network.neutron [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updated VIF entry in instance network info cache for port 2cd6eb89-f768-4ee6-93a2-386b83c42638. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 850.024129] env[65788]: DEBUG nova.network.neutron [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 850.136784] env[65788]: DEBUG nova.network.neutron [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 850.272977] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ebd4db-f51d-3598-45d0-ea3c0ce4788f, 'name': SearchDatastore_Task, 'duration_secs': 0.0132} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.273669] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.274144] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 850.274464] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-976dc110-289a-445b-b4f1-38b779447d0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.283053] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 850.283053] env[65788]: value = "task-4662448" [ 850.283053] env[65788]: _type = "Task" [ 850.283053] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.293645] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662448, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.506828] env[65788]: DEBUG nova.compute.manager [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 850.507145] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.508470] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617baa0d-f6d3-4fde-8031-d1150023d1d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.513237] env[65788]: INFO nova.compute.manager [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Swapping old allocation on dict_keys(['3a6417f0-b161-4086-8a0e-1cb139eae377']) held by migration 894f791d-94eb-4f0f-b93a-e53dccce48e2 for instance [ 850.521389] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.521724] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8af07f9f-387c-4e2d-97b7-4da841646611 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.528385] env[65788]: INFO nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 894f791d-94eb-4f0f-b93a-e53dccce48e2 has allocations against this compute host but is not found in the database. [ 850.528559] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f4a490a2-f2b1-4eac-8c1a-a18758583c70 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.528683] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f160e4a5-1e91-495e-800e-116ec435d8e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.528827] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 37471e59-1809-4df3-8c40-20921d04d18e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.528948] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Migration b5064230-67b0-4b10-b1f0-9c2db8d69267 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 850.529090] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f75000e6-f3bf-4951-a541-3aa185339054 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.529223] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance a35ced42-4317-49b4-b4cc-4ed7e2c85c64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.529291] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 3a06a833-5aaa-4b5d-88b3-8a1d469580af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 850.532924] env[65788]: DEBUG oslo_concurrency.lockutils [req-684f3f2c-12c6-4845-8cf2-20f37eccad3d req-c9458e3b-837f-4afc-976f-6516f34c68ad service nova] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.533492] env[65788]: DEBUG oslo_vmware.api [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 850.533492] env[65788]: value = "task-4662449" [ 850.533492] env[65788]: _type = "Task" [ 850.533492] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.538915] env[65788]: DEBUG nova.scheduler.client.report [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Overwriting current allocation {'allocations': {'3a6417f0-b161-4086-8a0e-1cb139eae377': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 83}}, 'project_id': 'a38ce4af9b414c778a69cd3f64f5f6bf', 'user_id': 'bedb4502cfc74d76b0c6df2d9173855e', 'consumer_generation': 1} on consumer f4a490a2-f2b1-4eac-8c1a-a18758583c70 {{(pid=65788) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 850.545404] env[65788]: DEBUG oslo_vmware.api [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.616969] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 850.642091] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.642930] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Instance network_info: |[{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 850.643503] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:9f:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ef746c57-cd18-4883-a0e9-c52937aaf41d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '27626e1a-51fe-4f5a-9b93-8ab4c1b8b694', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.652413] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 850.652413] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.652976] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f023036-fbb2-44ba-b7f9-ee3c08cf1fe0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.678774] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.679139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquired lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.679497] env[65788]: DEBUG nova.network.neutron [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 850.693394] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.693394] env[65788]: value = "task-4662450" [ 850.693394] env[65788]: _type = "Task" [ 850.693394] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.708468] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662450, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.799022] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662448, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.035542] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d042a386-0d16-4e2d-a8e7-7641c07bcb35 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 851.047966] env[65788]: DEBUG oslo_vmware.api [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662449, 'name': PowerOffVM_Task, 'duration_secs': 0.288747} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.048341] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 851.048536] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 851.048825] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d7826f7-a427-43bb-8faa-798bdc3dc5e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.150773] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 851.151097] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 851.151349] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleting the datastore file [datastore1] eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 851.151653] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f0fbe1d-cec8-4b9c-8a38-ee7226b841b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.160941] env[65788]: DEBUG oslo_vmware.api [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 851.160941] env[65788]: value = "task-4662452" [ 851.160941] env[65788]: _type = "Task" [ 851.160941] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.171550] env[65788]: DEBUG oslo_vmware.api [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.187180] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.187997] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.188467] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.206033] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662450, 'name': CreateVM_Task, 'duration_secs': 0.456645} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.206215] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.206599] env[65788]: WARNING neutronclient.v2_0.client [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.206990] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.208054] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 851.208054] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 851.208054] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2d0406e-ca68-40fe-9cb6-2e84f35fea97 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.214492] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 851.214492] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a024ac-0f3c-8759-0195-04b87440ead3" [ 851.214492] env[65788]: _type = "Task" [ 851.214492] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.227168] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a024ac-0f3c-8759-0195-04b87440ead3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.294828] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662448, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581227} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.295142] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 851.295366] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.295644] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-36f2b846-3684-414b-88b2-b1d45d8d7eff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.304591] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.305044] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.316082] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 851.316082] env[65788]: value = "task-4662453" [ 851.316082] env[65788]: _type = "Task" [ 851.316082] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.325822] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.377144] env[65788]: WARNING neutronclient.v2_0.client [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.377967] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.378345] env[65788]: WARNING openstack [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.472898] env[65788]: DEBUG nova.network.neutron [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [{"id": "04994181-be30-4c51-b4a1-ce198600a5c3", "address": "fa:16:3e:47:34:b4", "network": {"id": "a804ed9e-1eb6-4098-885a-b30fe8c9b9c6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.145", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "a305d1aa178f40f2a9e68e92449a002b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92233552-2c0c-416e-9bf3-bfcca8eda2dc", "external-id": "nsx-vlan-transportzone-251", "segmentation_id": 251, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04994181-be", "ovs_interfaceid": "04994181-be30-4c51-b4a1-ce198600a5c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 851.543163] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 851.673574] env[65788]: DEBUG oslo_vmware.api [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263003} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.673909] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.674240] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.674466] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.674649] env[65788]: INFO nova.compute.manager [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 1.17 seconds to destroy the instance on the hypervisor. [ 851.675019] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 851.675321] env[65788]: DEBUG nova.compute.manager [-] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 851.675409] env[65788]: DEBUG nova.network.neutron [-] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 851.675727] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.676403] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.676796] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.727682] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a024ac-0f3c-8759-0195-04b87440ead3, 'name': SearchDatastore_Task, 'duration_secs': 0.018385} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.728110] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.728405] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 851.728740] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.728948] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 851.729314] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 851.729926] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ed43dee-c905-4039-a0e9-7e47f8d6a2ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.741391] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 851.741617] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 851.742482] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ccd4b64-5893-4c2d-93a1-5ba207cf4127 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.746202] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 851.753538] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 851.753538] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52461a97-c55f-2812-f726-4a5671c56be4" [ 851.753538] env[65788]: _type = "Task" [ 851.753538] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.761624] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52461a97-c55f-2812-f726-4a5671c56be4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.827396] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091711} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.827683] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.828570] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921d7810-78ec-4c7c-b505-63653de4bef7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.852387] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.852839] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbb0db49-1c3f-47e3-bb83-827fb8169d70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.874955] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 851.874955] env[65788]: value = "task-4662454" [ 851.874955] env[65788]: _type = "Task" [ 851.874955] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.884424] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.976015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Releasing lock "refresh_cache-f4a490a2-f2b1-4eac-8c1a-a18758583c70" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.976517] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.976825] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b26ba38b-f627-4e08-bcf3-e6d960e105bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.985913] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 851.985913] env[65788]: value = "task-4662455" [ 851.985913] env[65788]: _type = "Task" [ 851.985913] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.994865] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662455, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.048150] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d2d614f8-15d9-47d6-9e66-8e1f8fcc18da has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 852.264952] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52461a97-c55f-2812-f726-4a5671c56be4, 'name': SearchDatastore_Task, 'duration_secs': 0.017129} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.266638] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cfe54b2-b13f-46b0-bff8-dc2554a967c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.274524] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 852.274524] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e27ee7-e184-1a05-b129-13e2dbb77afb" [ 852.274524] env[65788]: _type = "Task" [ 852.274524] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.286376] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e27ee7-e184-1a05-b129-13e2dbb77afb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.385931] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662454, 'name': ReconfigVM_Task, 'duration_secs': 0.452056} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.386823] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Reconfigured VM instance instance-0000003a to attach disk [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.387262] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-063d3d6e-04a5-4df9-9fb9-9f56054a9c06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.395547] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 852.395547] env[65788]: value = "task-4662456" [ 852.395547] env[65788]: _type = "Task" [ 852.395547] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.406184] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662456, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.500283] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662455, 'name': PowerOffVM_Task, 'duration_secs': 0.382758} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.500731] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.501733] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:47:39Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='edeb65e9-37d7-4c27-800f-7029f9e8ded0',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1773652449',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 852.502071] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.502345] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 852.502655] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.502891] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 852.503150] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 852.503519] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.503771] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 852.504052] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 852.504329] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 852.504640] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 852.513618] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94eb584b-ea1c-4e5d-b85b-e1638dc2586f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.533159] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 852.533159] env[65788]: value = "task-4662457" [ 852.533159] env[65788]: _type = "Task" [ 852.533159] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.543628] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662457, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.551660] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 852.753921] env[65788]: DEBUG nova.network.neutron [-] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 852.787773] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e27ee7-e184-1a05-b129-13e2dbb77afb, 'name': SearchDatastore_Task, 'duration_secs': 0.02536} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.788067] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.788420] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/3a06a833-5aaa-4b5d-88b3-8a1d469580af.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 852.788710] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ee4f675-09be-40ba-b364-ac36b24cf69c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.798049] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 852.798049] env[65788]: value = "task-4662458" [ 852.798049] env[65788]: _type = "Task" [ 852.798049] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.808667] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.906526] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662456, 'name': Rename_Task, 'duration_secs': 0.221843} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.906811] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.907084] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-170e4adf-8270-4cd8-b05d-a7e528da217e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.916569] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 852.916569] env[65788]: value = "task-4662459" [ 852.916569] env[65788]: _type = "Task" [ 852.916569] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.927119] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662459, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.046425] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662457, 'name': ReconfigVM_Task, 'duration_secs': 0.325347} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.047560] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31930143-a666-4d25-abb5-afa018e0852a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.069138] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 8d7a33e6-08fb-481e-8fac-fdf13b19aa21 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 853.071187] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:47:39Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='edeb65e9-37d7-4c27-800f-7029f9e8ded0',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1773652449',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 853.072032] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 853.072032] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 853.072032] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 853.072032] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 853.072323] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 853.072323] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.072467] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 853.072569] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 853.072735] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 853.072906] env[65788]: DEBUG nova.virt.hardware [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 853.074066] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c30c57e3-1921-4377-9bb7-b0716f1dd911 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.082383] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 853.082383] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b65279-4e41-b765-30c5-ec7595467552" [ 853.082383] env[65788]: _type = "Task" [ 853.082383] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.096908] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b65279-4e41-b765-30c5-ec7595467552, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.259127] env[65788]: INFO nova.compute.manager [-] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Took 1.58 seconds to deallocate network for instance. [ 853.313520] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662458, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.435607] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662459, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.575830] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 853.576397] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 853.576478] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4416MB phys_disk=100GB used_disk=20GB total_vcpus=48 used_vcpus=19 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '18', 'num_vm_active': '13', 'num_task_None': '13', 'num_os_type_None': '18', 'num_proj_a38ce4af9b414c778a69cd3f64f5f6bf': '2', 'io_workload': '2', 'num_proj_ffbe44d63c1d432e97849f15615329e1': '2', 'num_proj_2b70d3c8627449eaa6372ebe3bd90233': '1', 'num_proj_06b3dfec7c0144928b4a59dbd27569b7': '2', 'num_proj_c963c936cf20402d9a65f5e61a11020a': '1', 'num_vm_resized': '2', 'num_task_resize_reverting': '1', 'num_proj_c4cffe8dfc424459a90e316ac4ac905f': '3', 'num_vm_rescued': '1', 'num_proj_9674e2a5c86b48db8c865a50331ab846': '1', 'num_proj_41f69a1056bc454e83bd9561cf12c506': '1', 'num_task_image_uploading': '2', 'num_proj_1fb2fb1250354a1ba3e151b909c67667': '2', 'num_proj_6dcc0ac928ee42b9a926c98464c4ec95': '1', 'num_proj_3e1713c7002a413fb27180469fded83e': '1', 'num_vm_building': '2', 'num_task_spawning': '2', 'num_proj_5e7aa18dcc41403389518d700c2c8b8a': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 853.595894] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b65279-4e41-b765-30c5-ec7595467552, 'name': SearchDatastore_Task, 'duration_secs': 0.014787} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.601851] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfiguring VM instance instance-00000028 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 853.602302] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d818a017-8c04-43c9-b38e-4d8d1d444ed2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.626608] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 853.626608] env[65788]: value = "task-4662460" [ 853.626608] env[65788]: _type = "Task" [ 853.626608] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.637897] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.770429] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.814439] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.768179} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.814771] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/3a06a833-5aaa-4b5d-88b3-8a1d469580af.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 853.815065] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 853.815449] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76d53397-fa55-41fb-a52b-46cd4d4c3d33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.824949] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 853.824949] env[65788]: value = "task-4662461" [ 853.824949] env[65788]: _type = "Task" [ 853.824949] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.838996] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662461, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.929206] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e143ec-9f7c-8d04-6fad-7a85bfcb68c4/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 853.930518] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffebca00-b9d0-48df-bd3c-66e00716b5de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.941583] env[65788]: DEBUG oslo_vmware.api [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662459, 'name': PowerOnVM_Task, 'duration_secs': 0.708937} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.942735] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.942910] env[65788]: INFO nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Took 9.34 seconds to spawn the instance on the hypervisor. [ 853.943141] env[65788]: DEBUG nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 853.943886] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e143ec-9f7c-8d04-6fad-7a85bfcb68c4/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 853.943886] env[65788]: ERROR oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e143ec-9f7c-8d04-6fad-7a85bfcb68c4/disk-0.vmdk due to incomplete transfer. [ 853.947469] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2d3519-584b-4551-93b6-0d35ee43dea6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.950203] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0aa4abda-dac5-4d57-90c4-2d15b6ef710a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.961777] env[65788]: DEBUG oslo_vmware.rw_handles [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e143ec-9f7c-8d04-6fad-7a85bfcb68c4/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 853.962089] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Uploaded image e6f1a08c-b319-4a8d-9b85-bd96892a6aab to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 853.964928] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 853.965529] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-368d89f2-db35-4e6a-a465-0acd984b1c80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.977754] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 853.977754] env[65788]: value = "task-4662462" [ 853.977754] env[65788]: _type = "Task" [ 853.977754] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.988487] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662462, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.079173] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e197c9-d206-4456-b601-40f6738c3f4d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.089752] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882de8d0-671b-45af-96f1-da820d996793 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.965424] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81bf617-fe20-409e-afa1-5ea212921443 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.968492] env[65788]: DEBUG nova.compute.manager [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 854.968680] env[65788]: DEBUG nova.compute.manager [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing instance network info cache due to event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 854.968892] env[65788]: DEBUG oslo_concurrency.lockutils [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.969044] env[65788]: DEBUG oslo_concurrency.lockutils [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.969329] env[65788]: DEBUG nova.network.neutron [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 854.972581] env[65788]: INFO nova.compute.manager [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Took 56.32 seconds to build instance. [ 854.987465] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662461, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155837} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.987797] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662460, 'name': ReconfigVM_Task, 'duration_secs': 0.504335} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.987997] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662462, 'name': Destroy_Task, 'duration_secs': 0.447825} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.988793] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.990073] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602e8bb3-ab14-4f68-87f1-2a3981825592 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.994366] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfigured VM instance instance-00000028 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 854.995232] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Destroyed the VM [ 854.995232] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 854.995703] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e31fe7-6fbd-4ed6-a0cc-1c46f56977f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.999655] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b466b4-8eee-4510-a21f-e839f7966aad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.001919] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ad0cfb8a-4c22-4bb4-a947-58f089c9f05f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.027341] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/3a06a833-5aaa-4b5d-88b3-8a1d469580af.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.050662] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-554b4521-d7e0-463e-afe7-0d45602bc464 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.065539] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.067304] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 855.067304] env[65788]: value = "task-4662463" [ 855.067304] env[65788]: _type = "Task" [ 855.067304] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.075254] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.075750] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3e26ab4-ef76-4cb0-8133-cd34148179fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.096669] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 855.096669] env[65788]: value = "task-4662464" [ 855.096669] env[65788]: _type = "Task" [ 855.096669] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.104108] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662463, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.105799] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 855.105799] env[65788]: value = "task-4662465" [ 855.105799] env[65788]: _type = "Task" [ 855.105799] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.109220] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662464, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.116606] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.116841] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.126829] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662465, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.473843] env[65788]: WARNING neutronclient.v2_0.client [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.475080] env[65788]: WARNING openstack [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.475525] env[65788]: WARNING openstack [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.484533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-21953979-9ae6-4c6e-993a-4539459218c2 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.841s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 855.577052] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.594976] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662463, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.608964] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662464, 'name': ReconfigVM_Task, 'duration_secs': 0.462407} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.609978] env[65788]: WARNING openstack [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.610351] env[65788]: WARNING openstack [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.616726] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/3a06a833-5aaa-4b5d-88b3-8a1d469580af.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.617905] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c4b677e0-89b5-46cc-8941-8ef4331a0753 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.622872] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 855.636250] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662465, 'name': ReconfigVM_Task, 'duration_secs': 0.473037} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.637817] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Reconfigured VM instance instance-00000028 to attach disk [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70/f4a490a2-f2b1-4eac-8c1a-a18758583c70.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.638278] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 855.638278] env[65788]: value = "task-4662466" [ 855.638278] env[65788]: _type = "Task" [ 855.638278] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.639030] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5625d2-b134-4ead-8b25-5c0caab35003 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.671820] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662466, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.673300] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a63592-1043-4b2b-9c3d-107039386374 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.698949] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a4f0b0-7935-4cb7-9610-fcb949aa8694 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.725937] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cd89a9-9556-4eae-bbb3-8805cb00b7ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.734629] env[65788]: WARNING neutronclient.v2_0.client [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 855.734734] env[65788]: WARNING openstack [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 855.735157] env[65788]: WARNING openstack [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 855.747659] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.748020] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a326dd72-1247-487f-883b-62a625421c25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.757316] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 855.757316] env[65788]: value = "task-4662467" [ 855.757316] env[65788]: _type = "Task" [ 855.757316] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.770845] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662467, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.792179] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.792860] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.853420] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "7c88e3ef-82d2-46ac-a350-999de8091c01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 855.853420] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 855.900049] env[65788]: DEBUG nova.network.neutron [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updated VIF entry in instance network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 855.900328] env[65788]: DEBUG nova.network.neutron [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 856.085526] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 856.085953] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.130s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.086289] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 40.122s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.086638] env[65788]: DEBUG nova.objects.instance [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 856.089970] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.090290] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Cleaning up deleted instances {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11916}} [ 856.099980] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662463, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.156812] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662466, 'name': Rename_Task, 'duration_secs': 0.210634} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.158362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.158822] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.159651] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97dfdb36-7d3c-497b-848f-6b9d9d6e5699 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.169016] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 856.169016] env[65788]: value = "task-4662468" [ 856.169016] env[65788]: _type = "Task" [ 856.169016] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.179752] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662468, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.271180] env[65788]: DEBUG oslo_vmware.api [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662467, 'name': PowerOnVM_Task, 'duration_secs': 0.491258} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.271551] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.296467] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 856.356178] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 856.403279] env[65788]: DEBUG oslo_concurrency.lockutils [req-ba321f47-f223-4cf7-8fb2-db5a0c7be3dd req-4274b176-5807-43a8-9875-e010887a3ea1 service nova] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 856.591588] env[65788]: DEBUG oslo_vmware.api [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662463, 'name': RemoveSnapshot_Task, 'duration_secs': 1.243453} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.603735] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] There are 35 instances to clean {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11925}} [ 856.604514] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5c7a1693-62f2-454e-9406-0b4a132ebf25] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 856.611311] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 856.617053] env[65788]: INFO nova.compute.manager [None req-fdd19434-aaf2-4abf-9190-841a2606a083 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Took 17.38 seconds to snapshot the instance on the hypervisor. [ 856.640865] env[65788]: DEBUG nova.compute.manager [req-a3ab1561-013c-4df6-8fd9-9208af7865f5 req-19674d5c-2dee-4a9e-b0da-0aee6bfcb424 service nova] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Received event network-vif-deleted-d6e04a22-9f86-43c9-ac42-95191eaf454b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 856.684170] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662468, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.817803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.876770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.113920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3475b270-6e7b-4a73-82da-c0806d9ff57f tempest-ServersAdmin275Test-239756992 tempest-ServersAdmin275Test-239756992-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.027s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.115192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 39.738s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.116750] env[65788]: INFO nova.compute.claims [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.122872] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f3c53f67-85bf-4c18-9313-75eb90862f78] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 857.183211] env[65788]: DEBUG oslo_vmware.api [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662468, 'name': PowerOnVM_Task, 'duration_secs': 0.905497} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.183500] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.183785] env[65788]: INFO nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Took 9.37 seconds to spawn the instance on the hypervisor. [ 857.184031] env[65788]: DEBUG nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 857.184937] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1644e606-b9d7-4e7d-a0c5-5ee63890e8ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.285905] env[65788]: INFO nova.compute.manager [None req-10a097a8-a181-4581-9117-dda0d778f822 tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance to original state: 'active' [ 857.433048] env[65788]: DEBUG nova.objects.instance [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lazy-loading 'flavor' on Instance uuid 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.627936] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: cef78616-dcbd-4ccd-ab65-c9d52d8b8b2e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 857.704895] env[65788]: INFO nova.compute.manager [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Took 52.44 seconds to build instance. [ 857.938687] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.938870] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.939221] env[65788]: WARNING neutronclient.v2_0.client [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 857.939905] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 857.940286] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 858.132167] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 99f2cc13-a559-4379-8332-d432e8bac36b] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 858.209037] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2392569e-2a65-4a3e-9c03-d70004d602e6 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.957s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.236496] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5258f653-5850-bf0e-1398-b5752eaed473/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.237259] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec71e86-d743-4ee3-bba3-62d3b1fc42ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.244857] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5258f653-5850-bf0e-1398-b5752eaed473/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.245063] env[65788]: ERROR oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5258f653-5850-bf0e-1398-b5752eaed473/disk-0.vmdk due to incomplete transfer. [ 858.248600] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-85a70170-d4e4-4306-8b29-a5bcc01231f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.258762] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5258f653-5850-bf0e-1398-b5752eaed473/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 858.259053] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Uploaded image b9915b20-93e4-4ad5-851f-b711fc0ab2c2 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 858.262174] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 858.265790] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6bf6ebbc-7070-4eb2-95b5-21344b1beeeb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.274054] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 858.274054] env[65788]: value = "task-4662469" [ 858.274054] env[65788]: _type = "Task" [ 858.274054] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.284252] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662469, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.457240] env[65788]: DEBUG nova.network.neutron [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 858.636694] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 6be541bd-06fd-4959-9987-d250011b3f1d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 858.667509] env[65788]: DEBUG nova.compute.manager [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 858.667710] env[65788]: DEBUG nova.compute.manager [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing instance network info cache due to event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 858.667956] env[65788]: DEBUG oslo_concurrency.lockutils [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.668156] env[65788]: DEBUG oslo_concurrency.lockutils [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 858.669301] env[65788]: DEBUG nova.network.neutron [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 858.773916] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d5c4d7-527b-4297-b8b7-d19ff5fd8069 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.786650] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662469, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.789327] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876e741a-7a3e-4095-ace7-96f79baa06c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.826278] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3e7ea7-3c64-4585-b2d7-adbec2768b66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.836156] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eecf1b-36ed-42e1-a965-6d39ddcb7aef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.854124] env[65788]: DEBUG nova.compute.provider_tree [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.913225] env[65788]: DEBUG nova.compute.manager [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 858.913445] env[65788]: DEBUG nova.compute.manager [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing instance network info cache due to event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 858.913635] env[65788]: DEBUG oslo_concurrency.lockutils [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.962870] env[65788]: WARNING neutronclient.v2_0.client [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 858.962870] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 858.963255] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.139398] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 72111766-217d-469f-ad92-ca0ce0e67090] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 859.166396] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.166880] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.175210] env[65788]: WARNING neutronclient.v2_0.client [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 859.175849] env[65788]: WARNING openstack [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.176220] env[65788]: WARNING openstack [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.248240] env[65788]: WARNING neutronclient.v2_0.client [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 859.248932] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.249373] env[65788]: WARNING openstack [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.287759] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662469, 'name': Destroy_Task, 'duration_secs': 0.566794} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.288141] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Destroyed the VM [ 859.289025] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 859.289025] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cc8d22b0-cfed-4905-a762-9b3380cfcd48 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.302896] env[65788]: WARNING openstack [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.303371] env[65788]: WARNING openstack [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.312571] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 859.312571] env[65788]: value = "task-4662470" [ 859.312571] env[65788]: _type = "Task" [ 859.312571] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.322581] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662470, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.341837] env[65788]: DEBUG nova.network.neutron [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 859.357086] env[65788]: DEBUG nova.scheduler.client.report [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 859.372146] env[65788]: WARNING neutronclient.v2_0.client [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 859.372146] env[65788]: WARNING openstack [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 859.372146] env[65788]: WARNING openstack [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 859.499825] env[65788]: DEBUG nova.network.neutron [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updated VIF entry in instance network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 859.500259] env[65788]: DEBUG nova.network.neutron [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 859.615806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.616070] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.616694] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.616694] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.616694] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.620418] env[65788]: INFO nova.compute.manager [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Terminating instance [ 859.643518] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 3f20bd75-98ab-4798-834a-c0ffc3c6146d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 859.825032] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662470, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.846866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 859.847162] env[65788]: DEBUG nova.compute.manager [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Inject network info {{(pid=65788) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7777}} [ 859.847443] env[65788]: DEBUG nova.compute.manager [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] network_info to inject: |[{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7778}} [ 859.852743] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Reconfiguring VM instance to set the machine id {{(pid=65788) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 859.853157] env[65788]: DEBUG oslo_concurrency.lockutils [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.853482] env[65788]: DEBUG nova.network.neutron [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 859.855154] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82adae57-d41f-4f4a-900c-d5d900ffe2f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.868114] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.753s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.869184] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 859.871951] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.620s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.872240] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.874301] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 38.495s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.874606] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.877108] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.628s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.878025] env[65788]: DEBUG nova.objects.instance [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 859.886703] env[65788]: DEBUG oslo_vmware.api [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 859.886703] env[65788]: value = "task-4662471" [ 859.886703] env[65788]: _type = "Task" [ 859.886703] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.897537] env[65788]: DEBUG oslo_vmware.api [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662471, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.929740] env[65788]: INFO nova.scheduler.client.report [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocations for instance caceb296-5da2-4b4d-b955-5d7a238ff939 [ 859.931647] env[65788]: INFO nova.scheduler.client.report [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Deleted allocations for instance a813d102-1b77-4214-8eab-2cd66e002912 [ 860.005389] env[65788]: DEBUG oslo_concurrency.lockutils [req-5c34f7b5-5a6c-47b7-9ebf-03ac0be598ed req-786156ab-f4e4-431d-928b-14af2584649c service nova] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 860.124312] env[65788]: DEBUG nova.compute.manager [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 860.124576] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.126177] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055b3dd0-c98f-4d33-a19d-696470453aaa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.134801] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.135097] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41586d1e-1dbc-4298-a571-36671b1eadd1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.143207] env[65788]: DEBUG oslo_vmware.api [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 860.143207] env[65788]: value = "task-4662472" [ 860.143207] env[65788]: _type = "Task" [ 860.143207] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.147052] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 54ad3472-0cca-4dca-91f2-8c343f706926] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 860.156032] env[65788]: DEBUG oslo_vmware.api [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.180218] env[65788]: DEBUG nova.objects.instance [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lazy-loading 'flavor' on Instance uuid 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.325376] env[65788]: DEBUG oslo_vmware.api [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662470, 'name': RemoveSnapshot_Task, 'duration_secs': 0.912538} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.325656] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 860.325891] env[65788]: INFO nova.compute.manager [None req-b5e2f6f1-a7dd-4014-85b0-ed4f4e4e5b8e tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Took 16.25 seconds to snapshot the instance on the hypervisor. [ 860.368048] env[65788]: WARNING neutronclient.v2_0.client [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.369038] env[65788]: WARNING openstack [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.369539] env[65788]: WARNING openstack [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.382468] env[65788]: DEBUG nova.compute.utils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 860.383944] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 860.384166] env[65788]: DEBUG nova.network.neutron [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 860.384550] env[65788]: WARNING neutronclient.v2_0.client [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.384923] env[65788]: WARNING neutronclient.v2_0.client [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.385551] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.385931] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.409501] env[65788]: DEBUG oslo_vmware.api [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662471, 'name': ReconfigVM_Task, 'duration_secs': 0.183045} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.409726] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3297500f-fe7c-41b1-a58a-f0e7d5ff3ea7 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Reconfigured VM instance to set the machine id {{(pid=65788) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 860.446888] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5a9b23b4-afc6-4d9c-813e-4b4cce772667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "caceb296-5da2-4b4d-b955-5d7a238ff939" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.606s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.448633] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e86b76ed-5b93-4a02-b72b-1d4d6e0392f6 tempest-ServersAdmin275Test-968300172 tempest-ServersAdmin275Test-968300172-project-member] Lock "a813d102-1b77-4214-8eab-2cd66e002912" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.518s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.459571] env[65788]: DEBUG nova.policy [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e095550d08426cb59f11d2295b9a60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '246c0426008f478aa245d006c3c129eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 860.567904] env[65788]: WARNING openstack [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.568336] env[65788]: WARNING openstack [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.650456] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f604c16d-2a86-40d3-9891-5b33309b3047] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 860.655561] env[65788]: DEBUG oslo_vmware.api [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662472, 'name': PowerOffVM_Task, 'duration_secs': 0.327638} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.656077] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.656252] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.656506] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f81b9d39-98ae-4ec8-8851-4a192d269914 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.686662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.711279] env[65788]: WARNING neutronclient.v2_0.client [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 860.712594] env[65788]: WARNING openstack [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 860.712980] env[65788]: WARNING openstack [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 860.754782] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.755043] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.755239] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleting the datastore file [datastore2] f4a490a2-f2b1-4eac-8c1a-a18758583c70 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.755536] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1227abd-0599-4222-aa5d-481ab5149567 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.764697] env[65788]: DEBUG oslo_vmware.api [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 860.764697] env[65788]: value = "task-4662474" [ 860.764697] env[65788]: _type = "Task" [ 860.764697] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.780961] env[65788]: DEBUG oslo_vmware.api [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.804690] env[65788]: DEBUG nova.network.neutron [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Successfully created port: 48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 860.839116] env[65788]: DEBUG nova.network.neutron [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updated VIF entry in instance network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 860.839428] env[65788]: DEBUG nova.network.neutron [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 860.894759] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 860.899031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0fac6139-8bd1-4610-9da7-7c8decc18d24 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.903027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.746s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.903027] env[65788]: INFO nova.compute.claims [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.157437] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 6aa4b17b-816c-4d84-8f74-a81185f3af65] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 861.275391] env[65788]: DEBUG oslo_vmware.api [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183017} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.275845] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.276061] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.276254] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.276429] env[65788]: INFO nova.compute.manager [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Took 1.15 seconds to destroy the instance on the hypervisor. [ 861.276696] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 861.276897] env[65788]: DEBUG nova.compute.manager [-] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 861.276994] env[65788]: DEBUG nova.network.neutron [-] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 861.277263] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.277782] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 861.278053] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 861.344203] env[65788]: DEBUG oslo_concurrency.lockutils [req-5969e3b1-68fb-4b97-b836-960564e594d2 req-daab0a5f-4f6c-49e7-a048-a9459c7b0286 service nova] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.344708] env[65788]: DEBUG oslo_concurrency.lockutils [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.345329] env[65788]: WARNING neutronclient.v2_0.client [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.420378] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 861.420378] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.459588] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 861.660922] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 0548399d-ec7a-4d0c-b436-9ef93c4e2d37] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 861.921848] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 861.925030] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 861.963836] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 861.964180] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 861.964368] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 861.964565] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 861.964739] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 861.964896] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 861.965135] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.965368] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 861.965590] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 861.965781] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 861.965960] env[65788]: DEBUG nova.virt.hardware [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 861.967185] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335fb68e-9f5e-48e5-8286-c16eadcec3dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.981710] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5456409-8255-40cd-9717-61a74c9be0a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.165262] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 670f59c3-10b0-431f-a8b0-ef6f82a938be] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 862.235581] env[65788]: DEBUG nova.network.neutron [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 862.263250] env[65788]: DEBUG nova.network.neutron [-] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 862.441811] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cad7d1-38b6-4786-8273-77c350cb1ba6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.452036] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04508315-6fc0-45eb-be9d-cf96cdb34343 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.456528] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.484797] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f586b739-773a-46bd-9c5d-75a72f75d056 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.496164] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684001a4-b233-4033-b12e-3bea2e85fdf6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.547706] env[65788]: DEBUG nova.compute.provider_tree [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.547706] env[65788]: DEBUG nova.network.neutron [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Successfully updated port: 48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 862.592527] env[65788]: DEBUG nova.compute.manager [req-eab0bbd5-7d4f-47e8-8d3b-515c70fb360a req-a5e48873-d0c9-4c57-aa8f-2c76a1381c2a service nova] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Received event network-vif-deleted-04994181-be30-4c51-b4a1-ce198600a5c3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 862.671098] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 4b7f1c2f-2b1e-4d24-814b-c8095d875e3c] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 862.738850] env[65788]: WARNING neutronclient.v2_0.client [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.739750] env[65788]: WARNING openstack [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.740237] env[65788]: WARNING openstack [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.767423] env[65788]: INFO nova.compute.manager [-] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Took 1.49 seconds to deallocate network for instance. [ 862.869069] env[65788]: DEBUG nova.compute.manager [req-5fe65e63-ea22-4786-9386-1e053ab19e83 req-413ff4cd-9ac3-411f-a54e-a64599e812a4 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Received event network-vif-plugged-48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 862.869189] env[65788]: DEBUG oslo_concurrency.lockutils [req-5fe65e63-ea22-4786-9386-1e053ab19e83 req-413ff4cd-9ac3-411f-a54e-a64599e812a4 service nova] Acquiring lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.869405] env[65788]: DEBUG oslo_concurrency.lockutils [req-5fe65e63-ea22-4786-9386-1e053ab19e83 req-413ff4cd-9ac3-411f-a54e-a64599e812a4 service nova] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.869560] env[65788]: DEBUG oslo_concurrency.lockutils [req-5fe65e63-ea22-4786-9386-1e053ab19e83 req-413ff4cd-9ac3-411f-a54e-a64599e812a4 service nova] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.869847] env[65788]: DEBUG nova.compute.manager [req-5fe65e63-ea22-4786-9386-1e053ab19e83 req-413ff4cd-9ac3-411f-a54e-a64599e812a4 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] No waiting events found dispatching network-vif-plugged-48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 862.869901] env[65788]: WARNING nova.compute.manager [req-5fe65e63-ea22-4786-9386-1e053ab19e83 req-413ff4cd-9ac3-411f-a54e-a64599e812a4 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Received unexpected event network-vif-plugged-48d24f13-8cff-4392-b381-04ec0ce66f8d for instance with vm_state building and task_state spawning. [ 862.897015] env[65788]: WARNING openstack [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.897739] env[65788]: WARNING openstack [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.982644] env[65788]: WARNING neutronclient.v2_0.client [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.983517] env[65788]: WARNING openstack [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.984140] env[65788]: WARNING openstack [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.019013] env[65788]: DEBUG nova.scheduler.client.report [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 863.024886] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "refresh_cache-d042a386-0d16-4e2d-a8e7-7641c07bcb35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.025080] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "refresh_cache-d042a386-0d16-4e2d-a8e7-7641c07bcb35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.025607] env[65788]: DEBUG nova.network.neutron [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 863.097420] env[65788]: DEBUG nova.network.neutron [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 863.177595] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: bcdf4a61-a2e8-4ca6-92fc-faddf5371290] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 863.277557] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.524985] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.625s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.525600] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 863.529557] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.521s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.529807] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 863.532231] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.159s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.534194] env[65788]: INFO nova.compute.claims [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.537576] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.537956] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.571259] env[65788]: INFO nova.scheduler.client.report [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Deleted allocations for instance 93e1ade8-4c57-4db9-9da3-388e3eea91bd [ 863.585205] env[65788]: DEBUG nova.network.neutron [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 863.599553] env[65788]: DEBUG oslo_concurrency.lockutils [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.599815] env[65788]: DEBUG nova.compute.manager [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Inject network info {{(pid=65788) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7777}} [ 863.600098] env[65788]: DEBUG nova.compute.manager [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] network_info to inject: |[{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7778}} [ 863.605156] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Reconfiguring VM instance to set the machine id {{(pid=65788) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 863.605929] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1b1a8c4-edee-4936-a66c-d6012f2dabf1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.617501] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.617884] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.632922] env[65788]: DEBUG oslo_vmware.api [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 863.632922] env[65788]: value = "task-4662475" [ 863.632922] env[65788]: _type = "Task" [ 863.632922] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.641451] env[65788]: DEBUG oslo_vmware.api [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662475, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.681988] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 06f8fc6f-428c-4e40-8815-f0c78d27fa38] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 863.733495] env[65788]: WARNING neutronclient.v2_0.client [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 863.734196] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 863.734580] env[65788]: WARNING openstack [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 863.847316] env[65788]: DEBUG nova.network.neutron [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Updating instance_info_cache with network_info: [{"id": "48d24f13-8cff-4392-b381-04ec0ce66f8d", "address": "fa:16:3e:89:ea:96", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d24f13-8c", "ovs_interfaceid": "48d24f13-8cff-4392-b381-04ec0ce66f8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 864.054145] env[65788]: DEBUG nova.compute.utils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 864.060201] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 864.061138] env[65788]: DEBUG nova.network.neutron [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 864.061138] env[65788]: WARNING neutronclient.v2_0.client [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.061138] env[65788]: WARNING neutronclient.v2_0.client [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.061858] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 864.062384] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 864.078660] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12d21573-ed3b-47e2-86be-beeb1ac13c89 tempest-ServersListShow2100Test-399563236 tempest-ServersListShow2100Test-399563236-project-member] Lock "93e1ade8-4c57-4db9-9da3-388e3eea91bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.320s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.127184] env[65788]: DEBUG nova.policy [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '427df059789e4dfe9a78674f4ade2318', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d65617cb0224c18aa47c2b5440a88ed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 864.143932] env[65788]: DEBUG oslo_vmware.api [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662475, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.190776] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d240b3f7-d3e9-49e0-81a0-013bd1c2fe0c] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 864.352021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "refresh_cache-d042a386-0d16-4e2d-a8e7-7641c07bcb35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.352021] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Instance network_info: |[{"id": "48d24f13-8cff-4392-b381-04ec0ce66f8d", "address": "fa:16:3e:89:ea:96", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d24f13-8c", "ovs_interfaceid": "48d24f13-8cff-4392-b381-04ec0ce66f8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 864.352223] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:ea:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48d24f13-8cff-4392-b381-04ec0ce66f8d', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.360381] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Creating folder: Project (246c0426008f478aa245d006c3c129eb). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.361110] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cef00dc3-fdd9-41c8-ade2-3f7bde06b429 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.373858] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Created folder: Project (246c0426008f478aa245d006c3c129eb) in parent group-v910111. [ 864.373858] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Creating folder: Instances. Parent ref: group-v910293. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.374079] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c99c4dd-08bd-4578-9f85-719f80229968 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.384329] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Created folder: Instances in parent group-v910293. [ 864.384594] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 864.384801] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.385062] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-776adb9f-90bb-42cf-91c2-0977f20f487d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.405344] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.405344] env[65788]: value = "task-4662478" [ 864.405344] env[65788]: _type = "Task" [ 864.405344] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.416186] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662478, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.464144] env[65788]: DEBUG nova.network.neutron [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Successfully created port: 6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 864.560748] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 864.658572] env[65788]: DEBUG oslo_vmware.api [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662475, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.697423] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 12a0d54d-187c-49b7-ba13-b6c1ddf4e9b1] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 864.916068] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662478, 'name': CreateVM_Task, 'duration_secs': 0.445999} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.920198] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.921461] env[65788]: WARNING neutronclient.v2_0.client [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 864.922477] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.924018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.924018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 864.924018] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06da2b44-0d2e-4d84-827a-f0b420a10fbe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.932496] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 864.932496] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dd26de-75b3-9b75-468a-d61b67f3effc" [ 864.932496] env[65788]: _type = "Task" [ 864.932496] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.948417] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dd26de-75b3-9b75-468a-d61b67f3effc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.137902] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5539acfa-6301-4f43-bcd8-980f50323a8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.153627] env[65788]: DEBUG oslo_vmware.api [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662475, 'name': ReconfigVM_Task, 'duration_secs': 1.264298} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.160142] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-283b34cc-2d39-4444-9c19-bb0326bd0b84 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Reconfigured VM instance to set the machine id {{(pid=65788) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 865.160142] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9b4004-645e-4554-a9bf-e1931335f86b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.197873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df644bdc-ee4d-4c90-b39e-d4ae0ad13154 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.203073] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 42a9c6f2-08f0-4548-9101-685fc3930c7d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 865.207344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.207599] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.207801] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.207977] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.208167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 865.210310] env[65788]: INFO nova.compute.manager [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Terminating instance [ 865.212830] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ced2cb3-59f8-4f87-8a64-79f306a98a81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.232183] env[65788]: DEBUG nova.compute.provider_tree [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.373681] env[65788]: DEBUG nova.compute.manager [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.373834] env[65788]: DEBUG nova.compute.manager [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing instance network info cache due to event network-changed-a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 865.374157] env[65788]: DEBUG oslo_concurrency.lockutils [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Acquiring lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.374361] env[65788]: DEBUG oslo_concurrency.lockutils [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Acquired lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.374556] env[65788]: DEBUG nova.network.neutron [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Refreshing network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 865.445221] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dd26de-75b3-9b75-468a-d61b67f3effc, 'name': SearchDatastore_Task, 'duration_secs': 0.013032} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.445572] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.445816] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.446106] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.446256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.446435] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.446708] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad95fad1-f7ba-428a-b2d4-c5b4f534fa07 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.457429] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.457625] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.458372] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80ffe17b-702a-49ba-82cf-d32fe798b2fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.465571] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 865.465571] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed6bf0-f1f2-bb64-57cb-023f055d4166" [ 865.465571] env[65788]: _type = "Task" [ 865.465571] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.475309] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed6bf0-f1f2-bb64-57cb-023f055d4166, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.574532] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 865.605088] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 865.605439] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 865.605597] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 865.605780] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 865.605928] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 865.606092] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 865.606306] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.606460] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 865.606621] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 865.606781] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 865.606952] env[65788]: DEBUG nova.virt.hardware [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 865.607922] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ba25dd-40cf-498a-afee-f0235d01002c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.618341] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10bac28-fd4c-4400-8ade-f8246bdd189d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.674684] env[65788]: DEBUG nova.compute.manager [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Received event network-changed-48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 865.674896] env[65788]: DEBUG nova.compute.manager [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Refreshing instance network info cache due to event network-changed-48d24f13-8cff-4392-b381-04ec0ce66f8d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 865.675131] env[65788]: DEBUG oslo_concurrency.lockutils [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Acquiring lock "refresh_cache-d042a386-0d16-4e2d-a8e7-7641c07bcb35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.675276] env[65788]: DEBUG oslo_concurrency.lockutils [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Acquired lock "refresh_cache-d042a386-0d16-4e2d-a8e7-7641c07bcb35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.675435] env[65788]: DEBUG nova.network.neutron [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Refreshing network info cache for port 48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 865.712353] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2529ef4c-80d8-47b1-a73a-a02eea0ad8d8] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 865.720412] env[65788]: DEBUG nova.compute.manager [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 865.720635] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 865.721535] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc969c1-ec4d-431d-8176-b1eddd30338a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.731081] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.733016] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47618304-67ed-4249-92fe-aa4535bbee4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.736335] env[65788]: DEBUG nova.scheduler.client.report [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 865.743842] env[65788]: DEBUG oslo_vmware.api [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 865.743842] env[65788]: value = "task-4662479" [ 865.743842] env[65788]: _type = "Task" [ 865.743842] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.757903] env[65788]: DEBUG oslo_vmware.api [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.877719] env[65788]: WARNING neutronclient.v2_0.client [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 865.878435] env[65788]: WARNING openstack [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.878783] env[65788]: WARNING openstack [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.981784] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ed6bf0-f1f2-bb64-57cb-023f055d4166, 'name': SearchDatastore_Task, 'duration_secs': 0.010954} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.983414] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf67907f-a33c-4c47-9c67-7ae04493c650 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.992221] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 865.992221] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525224f4-2a36-1697-a9a4-a45b2c874386" [ 865.992221] env[65788]: _type = "Task" [ 865.992221] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.002357] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525224f4-2a36-1697-a9a4-a45b2c874386, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.029544] env[65788]: WARNING openstack [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.029958] env[65788]: WARNING openstack [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.178696] env[65788]: WARNING neutronclient.v2_0.client [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.178696] env[65788]: WARNING openstack [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.178982] env[65788]: WARNING openstack [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.191400] env[65788]: WARNING neutronclient.v2_0.client [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.192200] env[65788]: WARNING openstack [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.192543] env[65788]: WARNING openstack [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.209068] env[65788]: DEBUG nova.network.neutron [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Successfully updated port: 6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 866.215900] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 27ce0779-2fa2-44d2-a424-072183c022a6] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 866.243801] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.710s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.243801] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 866.250963] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.549s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.253132] env[65788]: INFO nova.compute.claims [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.274655] env[65788]: DEBUG oslo_vmware.api [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662479, 'name': PowerOffVM_Task, 'duration_secs': 0.207676} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.274655] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.274655] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.274655] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b80d9e57-a0c2-4176-b314-434a5c03b16d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.336758] env[65788]: DEBUG nova.network.neutron [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updated VIF entry in instance network info cache for port a52878da-edb4-4e48-8028-2ef5a48d8253. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 866.337154] env[65788]: DEBUG nova.network.neutron [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [{"id": "a52878da-edb4-4e48-8028-2ef5a48d8253", "address": "fa:16:3e:a5:e9:eb", "network": {"id": "f80bfd34-24c8-486d-a849-8c598c1755f8", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-177505214-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41f69a1056bc454e83bd9561cf12c506", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c791d09c-1086-4ee1-bcde-6ca7d259cabd", "external-id": "nsx-vlan-transportzone-990", "segmentation_id": 990, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa52878da-ed", "ovs_interfaceid": "a52878da-edb4-4e48-8028-2ef5a48d8253", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.350984] env[65788]: WARNING openstack [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.351487] env[65788]: WARNING openstack [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.360175] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.360407] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.360585] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Deleting the datastore file [datastore2] 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.361155] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0435daf2-2839-4408-a9bc-1cca6c2bf686 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.371030] env[65788]: DEBUG oslo_vmware.api [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for the task: (returnval){ [ 866.371030] env[65788]: value = "task-4662481" [ 866.371030] env[65788]: _type = "Task" [ 866.371030] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.379195] env[65788]: DEBUG oslo_vmware.api [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.389273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "bbf6a242-9873-44b9-8938-2f8df11a1018" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.389742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.390012] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "bbf6a242-9873-44b9-8938-2f8df11a1018-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 866.390217] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 866.390400] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 866.393636] env[65788]: INFO nova.compute.manager [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Terminating instance [ 866.421798] env[65788]: WARNING neutronclient.v2_0.client [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.422539] env[65788]: WARNING openstack [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.423049] env[65788]: WARNING openstack [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.502244] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525224f4-2a36-1697-a9a4-a45b2c874386, 'name': SearchDatastore_Task, 'duration_secs': 0.019518} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.502739] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.502985] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d042a386-0d16-4e2d-a8e7-7641c07bcb35/d042a386-0d16-4e2d-a8e7-7641c07bcb35.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.504128] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfb9fa63-9990-4e0d-9227-4fa051b1458c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.513332] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 866.513332] env[65788]: value = "task-4662482" [ 866.513332] env[65788]: _type = "Task" [ 866.513332] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.517913] env[65788]: DEBUG nova.network.neutron [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Updated VIF entry in instance network info cache for port 48d24f13-8cff-4392-b381-04ec0ce66f8d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 866.518538] env[65788]: DEBUG nova.network.neutron [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Updating instance_info_cache with network_info: [{"id": "48d24f13-8cff-4392-b381-04ec0ce66f8d", "address": "fa:16:3e:89:ea:96", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48d24f13-8c", "ovs_interfaceid": "48d24f13-8cff-4392-b381-04ec0ce66f8d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.530162] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.713454] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "refresh_cache-7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.713454] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquired lock "refresh_cache-7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.713609] env[65788]: DEBUG nova.network.neutron [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 866.721043] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 957f14fc-23f1-4334-b672-b3a75398f716] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 866.760402] env[65788]: DEBUG nova.compute.utils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 866.762160] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 866.763087] env[65788]: DEBUG nova.network.neutron [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 866.763477] env[65788]: WARNING neutronclient.v2_0.client [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.763885] env[65788]: WARNING neutronclient.v2_0.client [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.764469] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.764821] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.828109] env[65788]: DEBUG nova.policy [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd8168b496e84c939c0fe5602f00c6af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a84103bf9c94a6bbbb500854a3b9f9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 866.840450] env[65788]: DEBUG oslo_concurrency.lockutils [req-586dd4c6-ce1f-4435-8e6a-d120c1e6b992 req-a4470791-d244-4b58-8936-f77a10dc9a68 service nova] Releasing lock "refresh_cache-2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.884317] env[65788]: DEBUG oslo_vmware.api [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Task: {'id': task-4662481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145913} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.884603] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.885605] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.885605] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.886178] env[65788]: INFO nova.compute.manager [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Took 1.17 seconds to destroy the instance on the hypervisor. [ 866.886178] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 866.886340] env[65788]: DEBUG nova.compute.manager [-] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 866.886542] env[65788]: DEBUG nova.network.neutron [-] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 866.886895] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 866.887553] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.887819] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.899332] env[65788]: DEBUG nova.compute.manager [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 866.899434] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.901483] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de50b735-f829-4911-9d9d-10801d7c7c40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.911794] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.913637] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2eddbec-f8f4-497a-bcc2-162c477e6aeb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.924982] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 866.924982] env[65788]: value = "task-4662483" [ 866.924982] env[65788]: _type = "Task" [ 866.924982] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.937621] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662483, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.003690] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.027035] env[65788]: DEBUG oslo_concurrency.lockutils [req-3bedf0f3-1cfc-46c6-981f-9f61f6d8898a req-146b1899-ac41-4078-ac2a-85d26ee649fa service nova] Releasing lock "refresh_cache-d042a386-0d16-4e2d-a8e7-7641c07bcb35" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.036499] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662482, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519604} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.036695] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d042a386-0d16-4e2d-a8e7-7641c07bcb35/d042a386-0d16-4e2d-a8e7-7641c07bcb35.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.036895] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.037204] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e461d4f-7345-420c-b1e1-aa573da3bf2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.048599] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 867.048599] env[65788]: value = "task-4662484" [ 867.048599] env[65788]: _type = "Task" [ 867.048599] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.066794] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662484, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.206577] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "b4f157a7-350c-4fff-8509-44426714846c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.206833] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "b4f157a7-350c-4fff-8509-44426714846c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.217335] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.217768] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.227053] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 78b545b3-1f84-4ea8-b180-f9039f2ea4b1] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 867.276306] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 867.348280] env[65788]: DEBUG nova.network.neutron [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Successfully created port: 3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 867.439682] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662483, 'name': PowerOffVM_Task, 'duration_secs': 0.401458} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.441155] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.441155] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.441155] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2b2307c-1e51-440a-a3d9-041b1bd37f62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.446306] env[65788]: DEBUG nova.network.neutron [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 867.527248] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.527537] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.527723] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Deleting the datastore file [datastore2] bbf6a242-9873-44b9-8938-2f8df11a1018 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.528054] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1eb4ad01-1e53-45e5-95bf-505131d05e65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.536818] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for the task: (returnval){ [ 867.536818] env[65788]: value = "task-4662486" [ 867.536818] env[65788]: _type = "Task" [ 867.536818] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.550645] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.561847] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662484, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105377} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.563665] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.564867] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf9e19e-8219-4a4c-bed8-0383a06f7924 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.591894] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] d042a386-0d16-4e2d-a8e7-7641c07bcb35/d042a386-0d16-4e2d-a8e7-7641c07bcb35.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.597298] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbd03c75-98cd-49ba-9a4a-3dee77c65de4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.627478] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 867.627478] env[65788]: value = "task-4662487" [ 867.627478] env[65788]: _type = "Task" [ 867.627478] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.637163] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.637612] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.649854] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662487, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.730843] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 15aba949-c04c-4021-add8-2e9d4e4771db] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 867.734432] env[65788]: WARNING neutronclient.v2_0.client [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 867.735421] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.735830] env[65788]: WARNING openstack [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.797040] env[65788]: DEBUG nova.compute.manager [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Received event network-vif-plugged-6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 867.797275] env[65788]: DEBUG oslo_concurrency.lockutils [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Acquiring lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.797480] env[65788]: DEBUG oslo_concurrency.lockutils [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.797642] env[65788]: DEBUG oslo_concurrency.lockutils [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.797806] env[65788]: DEBUG nova.compute.manager [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] No waiting events found dispatching network-vif-plugged-6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 867.797968] env[65788]: WARNING nova.compute.manager [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Received unexpected event network-vif-plugged-6fe0678d-9d42-48ef-ace8-37c30ef0cc23 for instance with vm_state building and task_state spawning. [ 867.798268] env[65788]: DEBUG nova.compute.manager [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Received event network-changed-6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 867.798466] env[65788]: DEBUG nova.compute.manager [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Refreshing instance network info cache due to event network-changed-6fe0678d-9d42-48ef-ace8-37c30ef0cc23. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 867.798651] env[65788]: DEBUG oslo_concurrency.lockutils [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Acquiring lock "refresh_cache-7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.890552] env[65788]: DEBUG nova.network.neutron [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Updating instance_info_cache with network_info: [{"id": "6fe0678d-9d42-48ef-ace8-37c30ef0cc23", "address": "fa:16:3e:1f:d9:bf", "network": {"id": "86c194fc-c12e-42b9-8c79-cda34b51f1fa", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-550529353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d65617cb0224c18aa47c2b5440a88ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe0678d-9d", "ovs_interfaceid": "6fe0678d-9d42-48ef-ace8-37c30ef0cc23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 868.035343] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6477b354-d823-4eb7-9573-a660378f4aa6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.060024] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac8ab5a-9404-44a9-b7bb-9f4c67f7fa3a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.064211] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.065828] env[65788]: DEBUG nova.network.neutron [-] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 868.097459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a3c48c-6cd6-4a8e-b084-6779583b189d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.107609] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9182c1-cbfe-4c25-946d-3ffd0a67f091 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.128334] env[65788]: DEBUG nova.compute.provider_tree [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.140153] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662487, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.209498] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.209753] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.209958] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 868.210157] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 868.210321] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 868.214910] env[65788]: INFO nova.compute.manager [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Terminating instance [ 868.248352] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 54b0945e-c9f5-4726-928e-de1d0ded5aae] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 868.296661] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 868.324657] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 868.324657] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 868.324657] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 868.325114] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 868.325114] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 868.325114] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 868.325250] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.325399] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 868.325617] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 868.325790] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 868.326718] env[65788]: DEBUG nova.virt.hardware [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 868.326895] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86096b0-4add-408d-9f80-f80ec7f44588 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.339355] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d53710c-d867-45a4-a9ee-f849b2f2c837 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.395034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Releasing lock "refresh_cache-7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.395034] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Instance network_info: |[{"id": "6fe0678d-9d42-48ef-ace8-37c30ef0cc23", "address": "fa:16:3e:1f:d9:bf", "network": {"id": "86c194fc-c12e-42b9-8c79-cda34b51f1fa", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-550529353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d65617cb0224c18aa47c2b5440a88ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe0678d-9d", "ovs_interfaceid": "6fe0678d-9d42-48ef-ace8-37c30ef0cc23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 868.395245] env[65788]: DEBUG oslo_concurrency.lockutils [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Acquired lock "refresh_cache-7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 868.395245] env[65788]: DEBUG nova.network.neutron [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Refreshing network info cache for port 6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 868.396461] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:d9:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e59b364d-b7f6-499d-b7dc-82b8a819aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fe0678d-9d42-48ef-ace8-37c30ef0cc23', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.404088] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Creating folder: Project (7d65617cb0224c18aa47c2b5440a88ed). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 868.405837] env[65788]: WARNING neutronclient.v2_0.client [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.406555] env[65788]: WARNING openstack [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.406962] env[65788]: WARNING openstack [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.414971] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-375f97e2-3a56-4589-846b-c4b5320a5df8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.433114] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Created folder: Project (7d65617cb0224c18aa47c2b5440a88ed) in parent group-v910111. [ 868.433358] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Creating folder: Instances. Parent ref: group-v910296. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 868.433975] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d62e447f-36c8-4459-964d-07c39987afa9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.451891] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Created folder: Instances in parent group-v910296. [ 868.452396] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 868.452683] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 868.452995] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c3c751b-78b8-4087-aafa-990979f6ccae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.476409] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.476409] env[65788]: value = "task-4662490" [ 868.476409] env[65788]: _type = "Task" [ 868.476409] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.488762] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662490, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.557239] env[65788]: DEBUG oslo_vmware.api [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Task: {'id': task-4662486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.747281} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.557537] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.557758] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 868.557940] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 868.558129] env[65788]: INFO nova.compute.manager [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Took 1.66 seconds to destroy the instance on the hypervisor. [ 868.558419] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 868.558676] env[65788]: DEBUG nova.compute.manager [-] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 868.558837] env[65788]: DEBUG nova.network.neutron [-] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 868.559136] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.559939] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.560055] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.569065] env[65788]: INFO nova.compute.manager [-] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Took 1.68 seconds to deallocate network for instance. [ 868.631300] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.636462] env[65788]: DEBUG nova.scheduler.client.report [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 868.645998] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662487, 'name': ReconfigVM_Task, 'duration_secs': 0.942537} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.646546] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Reconfigured VM instance instance-0000003c to attach disk [datastore1] d042a386-0d16-4e2d-a8e7-7641c07bcb35/d042a386-0d16-4e2d-a8e7-7641c07bcb35.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.647423] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d5d21dc-0556-4523-8f58-0f6dbcd43173 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.660909] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 868.660909] env[65788]: value = "task-4662491" [ 868.660909] env[65788]: _type = "Task" [ 868.660909] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.666393] env[65788]: WARNING openstack [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 868.666794] env[65788]: WARNING openstack [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 868.685168] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662491, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.718750] env[65788]: DEBUG nova.compute.manager [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 868.718946] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.719701] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00626788-c234-4bde-af3f-e3b4105db0ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.730961] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.731556] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8926ac4-b4bc-4a8d-920a-0f5e7f70c079 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.740733] env[65788]: DEBUG oslo_vmware.api [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 868.740733] env[65788]: value = "task-4662492" [ 868.740733] env[65788]: _type = "Task" [ 868.740733] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.750704] env[65788]: DEBUG oslo_vmware.api [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662492, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.751266] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 77dcec96-50e5-4753-b3cb-c6aec377bce2] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 868.987438] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662490, 'name': CreateVM_Task, 'duration_secs': 0.469362} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.987438] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.988459] env[65788]: WARNING neutronclient.v2_0.client [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 868.988561] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.988640] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 868.988973] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 868.989401] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2b0f8cc-2bce-46ec-bc7e-63ebf9801363 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.995858] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 868.995858] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520a734d-a798-80cb-5efe-0cc6cfb95956" [ 868.995858] env[65788]: _type = "Task" [ 868.995858] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.007010] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520a734d-a798-80cb-5efe-0cc6cfb95956, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.071664] env[65788]: DEBUG nova.network.neutron [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Successfully updated port: 3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 869.073739] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.143850] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.894s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.144401] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 869.148614] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.926s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.150502] env[65788]: INFO nova.compute.claims [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.155961] env[65788]: WARNING neutronclient.v2_0.client [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.157630] env[65788]: WARNING openstack [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.157630] env[65788]: WARNING openstack [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.176866] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662491, 'name': Rename_Task, 'duration_secs': 0.229572} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.177193] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.177448] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e0efa9a-d502-409d-8536-a323e196c06c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.184984] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 869.184984] env[65788]: value = "task-4662493" [ 869.184984] env[65788]: _type = "Task" [ 869.184984] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.194237] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662493, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.243203] env[65788]: DEBUG nova.compute.manager [req-cdc1a606-70bd-4ee8-9219-ef5a889f610f req-ea394756-263b-456a-816a-4deb5679c1f7 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Received event network-vif-plugged-3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 869.243203] env[65788]: DEBUG oslo_concurrency.lockutils [req-cdc1a606-70bd-4ee8-9219-ef5a889f610f req-ea394756-263b-456a-816a-4deb5679c1f7 service nova] Acquiring lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.243203] env[65788]: DEBUG oslo_concurrency.lockutils [req-cdc1a606-70bd-4ee8-9219-ef5a889f610f req-ea394756-263b-456a-816a-4deb5679c1f7 service nova] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.243203] env[65788]: DEBUG oslo_concurrency.lockutils [req-cdc1a606-70bd-4ee8-9219-ef5a889f610f req-ea394756-263b-456a-816a-4deb5679c1f7 service nova] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.243971] env[65788]: DEBUG nova.compute.manager [req-cdc1a606-70bd-4ee8-9219-ef5a889f610f req-ea394756-263b-456a-816a-4deb5679c1f7 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] No waiting events found dispatching network-vif-plugged-3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 869.244158] env[65788]: WARNING nova.compute.manager [req-cdc1a606-70bd-4ee8-9219-ef5a889f610f req-ea394756-263b-456a-816a-4deb5679c1f7 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Received unexpected event network-vif-plugged-3a1140bb-e791-40d3-a08e-d01e1bfb79b6 for instance with vm_state building and task_state spawning. [ 869.258030] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: b5967423-89e1-44eb-b4de-22d09adb51a8] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 869.259549] env[65788]: DEBUG oslo_vmware.api [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662492, 'name': PowerOffVM_Task, 'duration_secs': 0.219949} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.261842] env[65788]: DEBUG nova.network.neutron [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Updated VIF entry in instance network info cache for port 6fe0678d-9d42-48ef-ace8-37c30ef0cc23. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 869.262225] env[65788]: DEBUG nova.network.neutron [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Updating instance_info_cache with network_info: [{"id": "6fe0678d-9d42-48ef-ace8-37c30ef0cc23", "address": "fa:16:3e:1f:d9:bf", "network": {"id": "86c194fc-c12e-42b9-8c79-cda34b51f1fa", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-550529353-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d65617cb0224c18aa47c2b5440a88ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe0678d-9d", "ovs_interfaceid": "6fe0678d-9d42-48ef-ace8-37c30ef0cc23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 869.263681] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.263860] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.264234] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78a6f5cb-65b5-4d6f-8bdd-3618c0d4410c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.341105] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.341352] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.341698] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Deleting the datastore file [datastore2] 47ff3955-5d9b-4d90-b8b1-276a0ce58a21 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.341817] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecc8b7d8-a79b-483f-977c-65cd94ed1e31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.349682] env[65788]: DEBUG oslo_vmware.api [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 869.349682] env[65788]: value = "task-4662495" [ 869.349682] env[65788]: _type = "Task" [ 869.349682] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.362139] env[65788]: DEBUG oslo_vmware.api [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.507680] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520a734d-a798-80cb-5efe-0cc6cfb95956, 'name': SearchDatastore_Task, 'duration_secs': 0.034078} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.507977] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 869.509392] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.509392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.509392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.509392] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.509392] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c95ec34d-5e4e-4515-8b1d-157cf0e069dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.520559] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.520797] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 869.521624] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f3a6bb-0c86-4709-a8a0-5bafee153583 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.528233] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 869.528233] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52da1564-c605-5514-8ec0-a87374b7fa61" [ 869.528233] env[65788]: _type = "Task" [ 869.528233] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.539174] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52da1564-c605-5514-8ec0-a87374b7fa61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.575973] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.575973] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.576114] env[65788]: DEBUG nova.network.neutron [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 869.600016] env[65788]: DEBUG nova.network.neutron [-] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 869.655642] env[65788]: DEBUG nova.compute.utils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 869.659737] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 869.660184] env[65788]: DEBUG nova.network.neutron [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 869.660532] env[65788]: WARNING neutronclient.v2_0.client [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.660869] env[65788]: WARNING neutronclient.v2_0.client [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.661472] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.661973] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.697481] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662493, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.712963] env[65788]: DEBUG nova.policy [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd8168b496e84c939c0fe5602f00c6af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a84103bf9c94a6bbbb500854a3b9f9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 869.761214] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 350f3c86-063c-4ed5-895b-fc621b2fa825] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 869.764174] env[65788]: DEBUG oslo_concurrency.lockutils [req-b3fd4b1b-51c8-4e40-9c09-2839092c7e9f req-210921bb-860c-4b5c-9bbc-88388651d699 service nova] Releasing lock "refresh_cache-7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 869.834095] env[65788]: DEBUG nova.compute.manager [req-1c2a5c24-a8bb-4f0f-b7d2-f81b8cfbbdef req-bb66bdee-e090-4cf7-8629-228d39741175 service nova] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Received event network-vif-deleted-a52878da-edb4-4e48-8028-2ef5a48d8253 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 869.834095] env[65788]: DEBUG nova.compute.manager [req-1c2a5c24-a8bb-4f0f-b7d2-f81b8cfbbdef req-bb66bdee-e090-4cf7-8629-228d39741175 service nova] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Received event network-vif-deleted-773392a6-7451-4cad-95dd-aba2407f0497 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 869.860857] env[65788]: DEBUG oslo_vmware.api [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662495, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.384204} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.861124] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.861308] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.861485] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.861860] env[65788]: INFO nova.compute.manager [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Took 1.14 seconds to destroy the instance on the hypervisor. [ 869.861941] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 869.862153] env[65788]: DEBUG nova.compute.manager [-] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 869.862256] env[65788]: DEBUG nova.network.neutron [-] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 869.862509] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 869.863131] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 869.863436] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 869.901106] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.004690] env[65788]: DEBUG nova.network.neutron [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Successfully created port: 02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 870.040468] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52da1564-c605-5514-8ec0-a87374b7fa61, 'name': SearchDatastore_Task, 'duration_secs': 0.010898} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.041320] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1866a9ba-9cfe-458a-ad36-1fc5e75c4d69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.049355] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 870.049355] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eea6db-03ff-c075-30f0-6915ff683541" [ 870.049355] env[65788]: _type = "Task" [ 870.049355] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.059326] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eea6db-03ff-c075-30f0-6915ff683541, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.079577] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.079992] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.104105] env[65788]: INFO nova.compute.manager [-] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Took 1.55 seconds to deallocate network for instance. [ 870.153061] env[65788]: DEBUG nova.network.neutron [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 870.160912] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 870.180485] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.180658] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.198966] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662493, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.265327] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 4d8f46cd-1c36-4e43-8110-66e9c991f28d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 870.347602] env[65788]: WARNING neutronclient.v2_0.client [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 870.348374] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 870.348706] env[65788]: WARNING openstack [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 870.485683] env[65788]: DEBUG nova.network.neutron [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Updating instance_info_cache with network_info: [{"id": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "address": "fa:16:3e:e3:e2:2c", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1140bb-e7", "ovs_interfaceid": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 870.562523] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eea6db-03ff-c075-30f0-6915ff683541, 'name': SearchDatastore_Task, 'duration_secs': 0.010217} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.563129] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 870.563129] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada/7fd30bbe-5797-4d0f-8d01-5c8a953f2ada.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 870.563385] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f1ad8cf-2834-4b16-a4a9-5b861805f543 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.575204] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 870.575204] env[65788]: value = "task-4662496" [ 870.575204] env[65788]: _type = "Task" [ 870.575204] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.584409] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662496, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.613899] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.621564] env[65788]: DEBUG nova.network.neutron [-] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 870.689366] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f24dfe4-f09d-4d0d-acb7-62a9088c3595 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.701677] env[65788]: DEBUG oslo_vmware.api [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662493, 'name': PowerOnVM_Task, 'duration_secs': 1.01581} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.703647] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.703824] env[65788]: INFO nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Took 8.78 seconds to spawn the instance on the hypervisor. [ 870.703978] env[65788]: DEBUG nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 870.704926] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15394c27-d96b-4cef-b1dd-4fd401508619 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.708217] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695abf43-68c9-4e75-b21d-207a7102654a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.745534] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ece09e1-21e7-42e3-86ab-a69d46484ecc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.754648] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb13c1c-03d5-4a03-8619-c6b84b0dcbed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.774668] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5b701040-025c-4246-ad54-f2cf478e998d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 870.776972] env[65788]: DEBUG nova.compute.provider_tree [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.989619] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 870.990362] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Instance network_info: |[{"id": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "address": "fa:16:3e:e3:e2:2c", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1140bb-e7", "ovs_interfaceid": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 870.990846] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:e2:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a1140bb-e791-40d3-a08e-d01e1bfb79b6', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.999165] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Creating folder: Project (7a84103bf9c94a6bbbb500854a3b9f9a). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.999509] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a234c97-d259-45c4-92ec-511d28f8b1e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.012716] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Created folder: Project (7a84103bf9c94a6bbbb500854a3b9f9a) in parent group-v910111. [ 871.012990] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Creating folder: Instances. Parent ref: group-v910299. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.013231] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-707a26b4-948c-47f9-a52d-456d8a46d0ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.025435] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Created folder: Instances in parent group-v910299. [ 871.025669] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 871.025899] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.026154] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e96742f-0ab7-4094-a9cb-78794afa1cc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.049534] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.049534] env[65788]: value = "task-4662499" [ 871.049534] env[65788]: _type = "Task" [ 871.049534] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.063142] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662499, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.088437] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662496, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.124742] env[65788]: INFO nova.compute.manager [-] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Took 1.26 seconds to deallocate network for instance. [ 871.177249] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 871.205991] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 871.206284] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 871.206439] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 871.206616] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 871.206758] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 871.206900] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 871.207119] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.207273] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 871.207437] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 871.207599] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 871.207764] env[65788]: DEBUG nova.virt.hardware [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 871.208689] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14110b7-69a5-4860-92fb-9cfdd504d2eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.217654] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98187d48-8420-49d0-806d-3a4bf74b1d01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.253530] env[65788]: INFO nova.compute.manager [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Took 53.91 seconds to build instance. [ 871.280157] env[65788]: DEBUG nova.scheduler.client.report [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.283740] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 141aefe8-1b95-4963-854d-da79ddf143f7] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 871.298337] env[65788]: DEBUG nova.compute.manager [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Received event network-changed-3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 871.298613] env[65788]: DEBUG nova.compute.manager [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Refreshing instance network info cache due to event network-changed-3a1140bb-e791-40d3-a08e-d01e1bfb79b6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 871.298854] env[65788]: DEBUG oslo_concurrency.lockutils [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Acquiring lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.299013] env[65788]: DEBUG oslo_concurrency.lockutils [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Acquired lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.299305] env[65788]: DEBUG nova.network.neutron [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Refreshing network info cache for port 3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 871.560444] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662499, 'name': CreateVM_Task, 'duration_secs': 0.425443} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.560788] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.561155] env[65788]: WARNING neutronclient.v2_0.client [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.561530] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.561682] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.562076] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 871.562347] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75a4aacb-9065-44db-a7b5-0b73c4ad2de0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.568036] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 871.568036] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52787fbd-0e0f-c06f-51a3-1b436d650006" [ 871.568036] env[65788]: _type = "Task" [ 871.568036] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.577298] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52787fbd-0e0f-c06f-51a3-1b436d650006, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.586629] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662496, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588798} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.586774] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada/7fd30bbe-5797-4d0f-8d01-5c8a953f2ada.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 871.587166] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 871.587344] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fff164e-7346-47b3-8172-ccdd6afa97ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.595303] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 871.595303] env[65788]: value = "task-4662500" [ 871.595303] env[65788]: _type = "Task" [ 871.595303] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.605488] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662500, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.627329] env[65788]: DEBUG nova.network.neutron [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Successfully updated port: 02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 871.632875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 871.756031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c89596ee-9d08-444c-880c-ff9873231cfb tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.416s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.787418] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.639s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.788064] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 871.794619] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 4bedb939-e86b-42bd-a490-a73086ecfd7d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 871.794619] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.368s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.794854] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.797408] env[65788]: DEBUG oslo_concurrency.lockutils [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.244s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.797785] env[65788]: DEBUG oslo_concurrency.lockutils [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.798126] env[65788]: INFO nova.compute.manager [None req-74d09825-20e9-4c11-b84e-465eb2f9599e tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Successfully reverted task state from rebuilding on failure for instance. [ 871.804794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.394s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.805071] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 871.807306] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.311s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 871.809227] env[65788]: INFO nova.compute.claims [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.815839] env[65788]: WARNING neutronclient.v2_0.client [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 871.815839] env[65788]: WARNING openstack [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.815839] env[65788]: WARNING openstack [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 871.838344] env[65788]: INFO nova.scheduler.client.report [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Deleted allocations for instance 1ef08e8f-744d-4939-9433-a4301cd31e77 [ 871.846022] env[65788]: INFO nova.scheduler.client.report [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted allocations for instance 86ac6437-511a-4d72-aff0-0325e2d633f6 [ 871.919986] env[65788]: WARNING openstack [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 871.920481] env[65788]: WARNING openstack [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.013272] env[65788]: WARNING neutronclient.v2_0.client [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 872.014054] env[65788]: WARNING openstack [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.014474] env[65788]: WARNING openstack [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.052902] env[65788]: DEBUG nova.compute.manager [req-69b4a05c-c443-4a2d-bff6-0607976fda5a req-a0a691f9-0c80-4deb-84ba-d1ebb9f0fa32 service nova] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Received event network-vif-deleted-4a920eec-6022-4912-93e4-46affc7b4d18 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 872.083541] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52787fbd-0e0f-c06f-51a3-1b436d650006, 'name': SearchDatastore_Task, 'duration_secs': 0.030407} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.083867] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.084121] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.084378] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.084517] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 872.084695] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.085892] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b24cfc16-f492-4dad-8e79-ebdfd5018cd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.100661] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.100865] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.102075] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4771203-b99b-4500-8de3-5e1bff85c80e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.115515] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662500, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.293165} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.116284] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.117135] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afca811-e3e5-4bf8-94fd-1e9d85ab0c2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.123755] env[65788]: DEBUG nova.network.neutron [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Updated VIF entry in instance network info cache for port 3a1140bb-e791-40d3-a08e-d01e1bfb79b6. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 872.124130] env[65788]: DEBUG nova.network.neutron [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Updating instance_info_cache with network_info: [{"id": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "address": "fa:16:3e:e3:e2:2c", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1140bb-e7", "ovs_interfaceid": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 872.126789] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 872.126789] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5238e481-e392-539b-cb1d-3a57c5fba35d" [ 872.126789] env[65788]: _type = "Task" [ 872.126789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.140996] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "refresh_cache-5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.141215] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "refresh_cache-5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 872.141372] env[65788]: DEBUG nova.network.neutron [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 872.154289] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada/7fd30bbe-5797-4d0f-8d01-5c8a953f2ada.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.156036] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6509182f-7431-47aa-8062-04c876a2910e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.175321] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5238e481-e392-539b-cb1d-3a57c5fba35d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.181994] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 872.181994] env[65788]: value = "task-4662501" [ 872.181994] env[65788]: _type = "Task" [ 872.181994] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.192809] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662501, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.259061] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 872.298054] env[65788]: DEBUG nova.compute.utils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 872.298054] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 872.298771] env[65788]: DEBUG nova.network.neutron [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 872.299265] env[65788]: WARNING neutronclient.v2_0.client [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 872.299719] env[65788]: WARNING neutronclient.v2_0.client [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 872.300480] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.300951] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.312582] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 872.317041] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: c06991cd-15b0-4d48-849a-140720ee07a0] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 872.355182] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b63f097-9d17-48c2-a4bf-a31b34794ec0 tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "86ac6437-511a-4d72-aff0-0325e2d633f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.437s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.357701] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e4cf0021-051c-44ac-8afb-8ae12bf004ec tempest-ServersAdminNegativeTestJSON-2089691366 tempest-ServersAdminNegativeTestJSON-2089691366-project-member] Lock "1ef08e8f-744d-4939-9433-a4301cd31e77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.474s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.363554] env[65788]: DEBUG nova.policy [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd8168b496e84c939c0fe5602f00c6af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a84103bf9c94a6bbbb500854a3b9f9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 872.632959] env[65788]: DEBUG oslo_concurrency.lockutils [req-05b65218-c77d-497c-82cd-c52cc4acce40 req-9ce1eeed-9cfc-40f2-af85-9135ac748132 service nova] Releasing lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.640409] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5238e481-e392-539b-cb1d-3a57c5fba35d, 'name': SearchDatastore_Task, 'duration_secs': 0.042397} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.641675] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc8f5c4a-528b-475b-bdb6-ecb2ee9ace64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.645027] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.645268] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.658267] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 872.658267] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5265fb48-420f-30bc-1b3a-45f792168b0d" [ 872.658267] env[65788]: _type = "Task" [ 872.658267] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.670424] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5265fb48-420f-30bc-1b3a-45f792168b0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.689495] env[65788]: DEBUG nova.network.neutron [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 872.699660] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662501, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.700913] env[65788]: DEBUG nova.network.neutron [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Successfully created port: 680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 872.710332] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.711026] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 872.792079] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.827190] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 98a2efec-cb20-4e95-9ed5-5b2057a48c39] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 872.862514] env[65788]: WARNING neutronclient.v2_0.client [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 872.864383] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 872.864383] env[65788]: WARNING openstack [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.014601] env[65788]: DEBUG nova.network.neutron [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Updating instance_info_cache with network_info: [{"id": "02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa", "address": "fa:16:3e:57:59:0b", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02faaa1a-1e", "ovs_interfaceid": "02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 873.177396] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5265fb48-420f-30bc-1b3a-45f792168b0d, 'name': SearchDatastore_Task, 'duration_secs': 0.01112} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.177727] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 873.178009] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] d2d614f8-15d9-47d6-9e66-8e1f8fcc18da/d2d614f8-15d9-47d6-9e66-8e1f8fcc18da.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.178331] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-544a5ee9-5167-4f7c-9efc-4556f16e0eb3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.191504] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 873.191504] env[65788]: value = "task-4662502" [ 873.191504] env[65788]: _type = "Task" [ 873.191504] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.200076] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662501, 'name': ReconfigVM_Task, 'duration_secs': 0.660608} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.200836] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada/7fd30bbe-5797-4d0f-8d01-5c8a953f2ada.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.201569] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14172647-7ae8-486d-85e4-57f28b6c1bb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.207985] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.217170] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 873.217170] env[65788]: value = "task-4662503" [ 873.217170] env[65788]: _type = "Task" [ 873.217170] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.227912] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662503, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.335679] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 873.337625] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 39127362-8daf-43ff-99cb-ddb54c6fe095] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 873.355406] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe2b7e8-3a16-41a7-8273-80f87ca367b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.365983] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06e8c12-d176-4d92-8ee6-813d3e5d192d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.372388] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 873.372698] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 873.372899] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 873.373192] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 873.373437] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 873.374027] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 873.374027] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.374027] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 873.374572] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 873.374572] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 873.374755] env[65788]: DEBUG nova.virt.hardware [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 873.377215] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de140273-19f6-4b7d-87bc-2c1d15b15623 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.411020] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b435022-afce-4a15-95d2-1339ea488265 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.418257] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbeb759-b059-4286-9f17-0e928db5051f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.427780] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb3ec6f-3d61-409e-aea5-92816ff77336 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.450940] env[65788]: DEBUG nova.compute.provider_tree [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.519253] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "refresh_cache-5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 873.519684] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Instance network_info: |[{"id": "02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa", "address": "fa:16:3e:57:59:0b", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02faaa1a-1e", "ovs_interfaceid": "02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 873.520748] env[65788]: DEBUG nova.compute.manager [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 873.521351] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:59:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.529287] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 873.529929] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ff2205-6084-4b61-8d4b-d7e5b812dd4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.533488] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.533727] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-92320c11-7382-45f9-8f35-27177fc62dec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.558667] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.558667] env[65788]: value = "task-4662504" [ 873.558667] env[65788]: _type = "Task" [ 873.558667] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.568673] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662504, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.572304] env[65788]: DEBUG nova.compute.manager [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Received event network-vif-plugged-02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 873.572523] env[65788]: DEBUG oslo_concurrency.lockutils [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Acquiring lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.572742] env[65788]: DEBUG oslo_concurrency.lockutils [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 873.572913] env[65788]: DEBUG oslo_concurrency.lockutils [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 873.573111] env[65788]: DEBUG nova.compute.manager [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] No waiting events found dispatching network-vif-plugged-02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 873.573328] env[65788]: WARNING nova.compute.manager [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Received unexpected event network-vif-plugged-02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa for instance with vm_state building and task_state spawning. [ 873.573456] env[65788]: DEBUG nova.compute.manager [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Received event network-changed-02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 873.573617] env[65788]: DEBUG nova.compute.manager [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Refreshing instance network info cache due to event network-changed-02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 873.573798] env[65788]: DEBUG oslo_concurrency.lockutils [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Acquiring lock "refresh_cache-5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.573928] env[65788]: DEBUG oslo_concurrency.lockutils [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Acquired lock "refresh_cache-5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.574097] env[65788]: DEBUG nova.network.neutron [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Refreshing network info cache for port 02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 873.704108] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662502, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.728599] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662503, 'name': Rename_Task, 'duration_secs': 0.276517} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.729051] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.729427] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cef8e3f1-d3de-4eaf-bf4f-40281343bb28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.737993] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 873.737993] env[65788]: value = "task-4662505" [ 873.737993] env[65788]: _type = "Task" [ 873.737993] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.749458] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.844296] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 48ac772b-3b3b-403e-814d-b5290b17fcad] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 873.954818] env[65788]: DEBUG nova.scheduler.client.report [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 874.061180] env[65788]: INFO nova.compute.manager [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] instance snapshotting [ 874.067491] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee81a51-88c4-436d-8a2c-0676979fb304 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.076325] env[65788]: WARNING neutronclient.v2_0.client [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.076993] env[65788]: WARNING openstack [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.077361] env[65788]: WARNING openstack [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.084220] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662504, 'name': CreateVM_Task, 'duration_secs': 0.460626} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.098433] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.099376] env[65788]: WARNING neutronclient.v2_0.client [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.100411] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.100411] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.100411] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 874.101082] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d3ff48a-a1f3-4adc-8221-837fe7363f8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.103782] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da3cbe5d-a70a-406d-b82b-3b16bb13b1bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.105495] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.105719] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.105914] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.106104] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.106268] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.108873] env[65788]: INFO nova.compute.manager [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Terminating instance [ 874.117942] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 874.117942] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271efe2-c446-70e2-ac2a-594eb412b37a" [ 874.117942] env[65788]: _type = "Task" [ 874.117942] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.127268] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271efe2-c446-70e2-ac2a-594eb412b37a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.204324] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.829233} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.204589] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] d2d614f8-15d9-47d6-9e66-8e1f8fcc18da/d2d614f8-15d9-47d6-9e66-8e1f8fcc18da.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.204814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.205140] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13abd2d7-0140-4fa1-8df3-1b9be3319ee2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.212681] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 874.212681] env[65788]: value = "task-4662506" [ 874.212681] env[65788]: _type = "Task" [ 874.212681] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.221633] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662506, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.252311] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.347982] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 874.348286] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Cleaning up deleted instances with incomplete migration {{(pid=65788) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11954}} [ 874.375065] env[65788]: DEBUG nova.network.neutron [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Successfully updated port: 680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 874.431443] env[65788]: WARNING openstack [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.431849] env[65788]: WARNING openstack [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.460783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.653s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.461358] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 874.465565] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.888s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.465565] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.467940] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 24.720s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.527751] env[65788]: DEBUG nova.compute.manager [req-2a69bae5-8068-4f7d-a441-3e1931bcc4bc req-fbdfd692-12a6-4eb3-9196-bf7395402162 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Received event network-vif-plugged-680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 874.528911] env[65788]: DEBUG oslo_concurrency.lockutils [req-2a69bae5-8068-4f7d-a441-3e1931bcc4bc req-fbdfd692-12a6-4eb3-9196-bf7395402162 service nova] Acquiring lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.528911] env[65788]: DEBUG oslo_concurrency.lockutils [req-2a69bae5-8068-4f7d-a441-3e1931bcc4bc req-fbdfd692-12a6-4eb3-9196-bf7395402162 service nova] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.529046] env[65788]: DEBUG oslo_concurrency.lockutils [req-2a69bae5-8068-4f7d-a441-3e1931bcc4bc req-fbdfd692-12a6-4eb3-9196-bf7395402162 service nova] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.529514] env[65788]: DEBUG nova.compute.manager [req-2a69bae5-8068-4f7d-a441-3e1931bcc4bc req-fbdfd692-12a6-4eb3-9196-bf7395402162 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] No waiting events found dispatching network-vif-plugged-680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 874.529701] env[65788]: WARNING nova.compute.manager [req-2a69bae5-8068-4f7d-a441-3e1931bcc4bc req-fbdfd692-12a6-4eb3-9196-bf7395402162 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Received unexpected event network-vif-plugged-680d08e3-6217-4e78-82fa-ea979448f95c for instance with vm_state building and task_state spawning. [ 874.610552] env[65788]: WARNING neutronclient.v2_0.client [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.611263] env[65788]: WARNING openstack [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.611717] env[65788]: WARNING openstack [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.621024] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 874.621626] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-61def9a3-be29-4685-bc03-5b43f36f7a1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.627033] env[65788]: DEBUG nova.compute.manager [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 874.627261] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.627956] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c368930d-2fc6-463a-aad6-5ef629944c9a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.641444] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271efe2-c446-70e2-ac2a-594eb412b37a, 'name': SearchDatastore_Task, 'duration_secs': 0.01446} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.645073] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.645346] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.645582] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.645729] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.646053] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.646259] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.646522] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 874.646522] env[65788]: value = "task-4662507" [ 874.646522] env[65788]: _type = "Task" [ 874.646522] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.646735] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eefae5ef-33b2-41b6-94c9-80669a93b13c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.648643] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a1ba72e8-67f4-4cf7-b6bb-5284ef88c081 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.659271] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662507, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.660966] env[65788]: DEBUG oslo_vmware.api [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 874.660966] env[65788]: value = "task-4662508" [ 874.660966] env[65788]: _type = "Task" [ 874.660966] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.665700] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.665890] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.667147] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0067ee38-b04f-4247-8c71-23491608ef93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.672932] env[65788]: DEBUG oslo_vmware.api [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.681720] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 874.681720] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522f8c4d-e31d-5a9c-a1d6-164b2ef861d8" [ 874.681720] env[65788]: _type = "Task" [ 874.681720] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.691526] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522f8c4d-e31d-5a9c-a1d6-164b2ef861d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.726502] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662506, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.307448} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.726870] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.728391] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f7efa0-cede-4df0-8a1c-2d7c15ac18a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.755082] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] d2d614f8-15d9-47d6-9e66-8e1f8fcc18da/d2d614f8-15d9-47d6-9e66-8e1f8fcc18da.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.757937] env[65788]: DEBUG nova.network.neutron [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Updated VIF entry in instance network info cache for port 02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 874.758333] env[65788]: DEBUG nova.network.neutron [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Updating instance_info_cache with network_info: [{"id": "02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa", "address": "fa:16:3e:57:59:0b", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02faaa1a-1e", "ovs_interfaceid": "02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 874.763816] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d6fb643-484e-46bc-a974-c613fe7b4f99 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.783255] env[65788]: DEBUG oslo_concurrency.lockutils [req-c9773600-523d-4fb1-b313-650bc5413a70 req-abb91c30-2b5d-43fa-b459-37361a823608 service nova] Releasing lock "refresh_cache-5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.791519] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662505, 'name': PowerOnVM_Task} progress is 90%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.793098] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 874.793098] env[65788]: value = "task-4662509" [ 874.793098] env[65788]: _type = "Task" [ 874.793098] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.850492] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 874.877984] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "refresh_cache-8d7a33e6-08fb-481e-8fac-fdf13b19aa21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.878255] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "refresh_cache-8d7a33e6-08fb-481e-8fac-fdf13b19aa21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 874.878561] env[65788]: DEBUG nova.network.neutron [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 874.972739] env[65788]: DEBUG nova.compute.utils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 874.977457] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 874.977696] env[65788]: DEBUG nova.network.neutron [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 874.978126] env[65788]: WARNING neutronclient.v2_0.client [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.978376] env[65788]: WARNING neutronclient.v2_0.client [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 874.978983] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.979367] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.001525] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a540691-f78e-4243-90cd-2cabf37e61fb tempest-ServerActionsV293TestJSON-1619444112 tempest-ServerActionsV293TestJSON-1619444112-project-member] Lock "263cd2fb-7eeb-4395-b1e9-b829940ebd35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.829s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 875.079270] env[65788]: DEBUG nova.policy [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b85cdc0332f4f609d5021f5ad462511', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a93220fe73e941a58f8df639e3764fd7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 875.165384] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662507, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.174510] env[65788]: DEBUG oslo_vmware.api [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662508, 'name': PowerOffVM_Task, 'duration_secs': 0.317875} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.174787] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.175065] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.175258] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1f2edfe-c44d-4fe3-b294-6dcc47e2348f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.192965] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522f8c4d-e31d-5a9c-a1d6-164b2ef861d8, 'name': SearchDatastore_Task, 'duration_secs': 0.045459} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.193828] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef195ded-57c3-4b3a-acf0-1248e33978e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.201677] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 875.201677] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b37453-056d-c2a5-3ab6-5a4d1a634f97" [ 875.201677] env[65788]: _type = "Task" [ 875.201677] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.211890] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b37453-056d-c2a5-3ab6-5a4d1a634f97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.257122] env[65788]: DEBUG oslo_vmware.api [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662505, 'name': PowerOnVM_Task, 'duration_secs': 1.491816} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.257483] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 875.257860] env[65788]: INFO nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Took 9.68 seconds to spawn the instance on the hypervisor. [ 875.258955] env[65788]: DEBUG nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 875.258955] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1217945-1679-47ec-b7ab-088a8b82ce23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.279972] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.280519] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.280630] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleting the datastore file [datastore1] d3e2b5f6-20f1-40ac-b92e-e110ee119178 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.281149] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54f93369-4139-4a0a-b8bf-a70b9f8c4d0d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.289662] env[65788]: DEBUG oslo_vmware.api [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 875.289662] env[65788]: value = "task-4662511" [ 875.289662] env[65788]: _type = "Task" [ 875.289662] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.310894] env[65788]: DEBUG oslo_vmware.api [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662511, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.317824] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662509, 'name': ReconfigVM_Task, 'duration_secs': 0.315483} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.318457] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Reconfigured VM instance instance-0000003e to attach disk [datastore2] d2d614f8-15d9-47d6-9e66-8e1f8fcc18da/d2d614f8-15d9-47d6-9e66-8e1f8fcc18da.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.319342] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fb17f54-37e4-418e-b611-e93659785611 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.328014] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 875.328014] env[65788]: value = "task-4662512" [ 875.328014] env[65788]: _type = "Task" [ 875.328014] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.343595] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662512, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.384236] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.384236] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.478626] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 875.491109] env[65788]: DEBUG nova.network.neutron [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 875.517785] env[65788]: DEBUG nova.network.neutron [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Successfully created port: 421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 875.551527] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.551921] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.640489] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35779890-dba4-443b-830f-c1ce1c35b637 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.649052] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f4bef7-84b4-4d78-a822-a7dee573b3a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.661666] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662507, 'name': CreateSnapshot_Task, 'duration_secs': 0.940025} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.687631] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 875.688696] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c464fd2e-cd14-4055-a545-78591d1fa3b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.691784] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2a41d1-713f-442e-b73c-1603c590983d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.703560] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf7440-a042-461e-9977-e99cc2ce1afd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.719677] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b37453-056d-c2a5-3ab6-5a4d1a634f97, 'name': SearchDatastore_Task, 'duration_secs': 0.012196} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.728090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 875.728427] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a/5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.728942] env[65788]: DEBUG nova.compute.provider_tree [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.730435] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d2b804c-f977-4ff4-b11f-33a73c000ad9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.741028] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 875.741028] env[65788]: value = "task-4662513" [ 875.741028] env[65788]: _type = "Task" [ 875.741028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.750417] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.784278] env[65788]: INFO nova.compute.manager [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Took 46.66 seconds to build instance. [ 875.800572] env[65788]: DEBUG oslo_vmware.api [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662511, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230922} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.800572] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.800716] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.800915] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.801142] env[65788]: INFO nova.compute.manager [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Took 1.17 seconds to destroy the instance on the hypervisor. [ 875.801714] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 875.802191] env[65788]: DEBUG nova.compute.manager [-] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 875.802293] env[65788]: DEBUG nova.network.neutron [-] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 875.802538] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 875.803092] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 875.803408] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 875.841029] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662512, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.221727] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 876.222084] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1b718980-2e31-495d-a0b4-d57175199f2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.232762] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 876.232762] env[65788]: value = "task-4662514" [ 876.232762] env[65788]: _type = "Task" [ 876.232762] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.233784] env[65788]: DEBUG nova.scheduler.client.report [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 876.250885] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662514, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.262778] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662513, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.286814] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9edba7f0-911a-4e87-9e65-2e55270b64a2 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.187s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.342330] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662512, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.466339] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.490637] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 876.503263] env[65788]: WARNING neutronclient.v2_0.client [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 876.503263] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.503542] env[65788]: WARNING openstack [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.531466] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 876.531723] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.531882] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 876.532518] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.532518] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 876.532622] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 876.534710] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.534710] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 876.534710] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 876.534710] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 876.534710] env[65788]: DEBUG nova.virt.hardware [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 876.535176] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02fff6b-d8bd-4253-9b59-a41dcfcfa05a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.545681] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e12f2b1-2812-4933-82a1-87bfb12ec69f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.749325] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662514, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.761234] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.789125} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.763501] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a/5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.763879] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.764701] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-913b5496-a7ec-4abd-acaf-f358a5ac8b08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.775038] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 876.775038] env[65788]: value = "task-4662515" [ 876.775038] env[65788]: _type = "Task" [ 876.775038] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.787033] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662515, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.845553] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662512, 'name': Rename_Task, 'duration_secs': 1.158497} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.845553] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.845748] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-640619c8-9836-4abe-84bd-e9327df76e88 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.854590] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 876.854590] env[65788]: value = "task-4662516" [ 876.854590] env[65788]: _type = "Task" [ 876.854590] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.863779] env[65788]: DEBUG nova.network.neutron [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Updating instance_info_cache with network_info: [{"id": "680d08e3-6217-4e78-82fa-ea979448f95c", "address": "fa:16:3e:d6:70:e1", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap680d08e3-62", "ovs_interfaceid": "680d08e3-6217-4e78-82fa-ea979448f95c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 876.868707] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662516, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.152184] env[65788]: DEBUG nova.network.neutron [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Successfully updated port: 421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 877.232866] env[65788]: DEBUG nova.network.neutron [-] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 877.245738] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662514, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.252023] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.783s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.254660] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.484s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.255118] env[65788]: DEBUG nova.objects.instance [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'resources' on Instance uuid eceafff8-7d28-4b9b-ade6-5736d5977efa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.294296] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662515, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080305} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.294618] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.295498] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970e1c73-9505-41d7-8a0e-34c5c1e00d83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.324190] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a/5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.324931] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89c71f18-42f8-4ff8-badd-d0fbf28cf1c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.354159] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 877.354159] env[65788]: value = "task-4662517" [ 877.354159] env[65788]: _type = "Task" [ 877.354159] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.366116] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662517, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.370846] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "refresh_cache-8d7a33e6-08fb-481e-8fac-fdf13b19aa21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.371262] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Instance network_info: |[{"id": "680d08e3-6217-4e78-82fa-ea979448f95c", "address": "fa:16:3e:d6:70:e1", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap680d08e3-62", "ovs_interfaceid": "680d08e3-6217-4e78-82fa-ea979448f95c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 877.371629] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662516, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.372575] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:70:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '680d08e3-6217-4e78-82fa-ea979448f95c', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.379846] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 877.380156] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.380462] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04b6339c-8a6f-4822-8f97-7c3563a098da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.401126] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.401126] env[65788]: value = "task-4662518" [ 877.401126] env[65788]: _type = "Task" [ 877.401126] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.410137] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662518, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.457220] env[65788]: DEBUG nova.compute.manager [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Received event network-changed-680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 877.457344] env[65788]: DEBUG nova.compute.manager [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Refreshing instance network info cache due to event network-changed-680d08e3-6217-4e78-82fa-ea979448f95c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 877.457727] env[65788]: DEBUG oslo_concurrency.lockutils [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Acquiring lock "refresh_cache-8d7a33e6-08fb-481e-8fac-fdf13b19aa21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.457727] env[65788]: DEBUG oslo_concurrency.lockutils [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Acquired lock "refresh_cache-8d7a33e6-08fb-481e-8fac-fdf13b19aa21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.457832] env[65788]: DEBUG nova.network.neutron [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Refreshing network info cache for port 680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 877.545206] env[65788]: DEBUG nova.compute.manager [req-86c4e94e-83a6-44e6-aec9-188483ab1345 req-cb0f7b6a-11d9-44f1-8a68-cac4234c97e8 service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Received event network-vif-plugged-421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 877.545421] env[65788]: DEBUG oslo_concurrency.lockutils [req-86c4e94e-83a6-44e6-aec9-188483ab1345 req-cb0f7b6a-11d9-44f1-8a68-cac4234c97e8 service nova] Acquiring lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 877.545659] env[65788]: DEBUG oslo_concurrency.lockutils [req-86c4e94e-83a6-44e6-aec9-188483ab1345 req-cb0f7b6a-11d9-44f1-8a68-cac4234c97e8 service nova] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 877.545785] env[65788]: DEBUG oslo_concurrency.lockutils [req-86c4e94e-83a6-44e6-aec9-188483ab1345 req-cb0f7b6a-11d9-44f1-8a68-cac4234c97e8 service nova] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.545936] env[65788]: DEBUG nova.compute.manager [req-86c4e94e-83a6-44e6-aec9-188483ab1345 req-cb0f7b6a-11d9-44f1-8a68-cac4234c97e8 service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] No waiting events found dispatching network-vif-plugged-421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 877.546120] env[65788]: WARNING nova.compute.manager [req-86c4e94e-83a6-44e6-aec9-188483ab1345 req-cb0f7b6a-11d9-44f1-8a68-cac4234c97e8 service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Received unexpected event network-vif-plugged-421c2150-8fa1-4607-a700-cba48466f00a for instance with vm_state building and task_state spawning. [ 877.656418] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "refresh_cache-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.656418] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquired lock "refresh_cache-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.656418] env[65788]: DEBUG nova.network.neutron [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 877.739667] env[65788]: INFO nova.compute.manager [-] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Took 1.94 seconds to deallocate network for instance. [ 877.752141] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662514, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.871866] env[65788]: INFO nova.scheduler.client.report [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocation for migration b5064230-67b0-4b10-b1f0-9c2db8d69267 [ 877.880052] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662517, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.893139] env[65788]: DEBUG oslo_vmware.api [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662516, 'name': PowerOnVM_Task, 'duration_secs': 0.543442} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.893577] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.893577] env[65788]: INFO nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Took 9.60 seconds to spawn the instance on the hypervisor. [ 877.893717] env[65788]: DEBUG nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 877.894699] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b2daf2-1297-44ee-a510-46189dd90a36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.920630] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662518, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.961916] env[65788]: WARNING neutronclient.v2_0.client [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 877.963145] env[65788]: WARNING openstack [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.963539] env[65788]: WARNING openstack [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.118261] env[65788]: WARNING openstack [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.118992] env[65788]: WARNING openstack [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.163135] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.163634] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.214529] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.214875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.215214] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.215685] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.215753] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.218594] env[65788]: DEBUG nova.network.neutron [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 878.223928] env[65788]: INFO nova.compute.manager [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Terminating instance [ 878.232512] env[65788]: WARNING neutronclient.v2_0.client [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.233215] env[65788]: WARNING openstack [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.233573] env[65788]: WARNING openstack [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.253405] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.253797] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.261808] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.269254] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662514, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.330893] env[65788]: WARNING neutronclient.v2_0.client [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.332144] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 878.332613] env[65788]: WARNING openstack [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 878.370545] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662517, 'name': ReconfigVM_Task, 'duration_secs': 0.836524} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.373647] env[65788]: DEBUG nova.network.neutron [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Updated VIF entry in instance network info cache for port 680d08e3-6217-4e78-82fa-ea979448f95c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 878.373647] env[65788]: DEBUG nova.network.neutron [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Updating instance_info_cache with network_info: [{"id": "680d08e3-6217-4e78-82fa-ea979448f95c", "address": "fa:16:3e:d6:70:e1", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap680d08e3-62", "ovs_interfaceid": "680d08e3-6217-4e78-82fa-ea979448f95c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 878.374986] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a/5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.375845] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0855bc2-50ee-4040-b1c0-5282f36fe9d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.387526] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 878.387526] env[65788]: value = "task-4662519" [ 878.387526] env[65788]: _type = "Task" [ 878.387526] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.395077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a0cb32e-ea89-4964-84ee-17a3e0e085c2 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 32.242s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.397634] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc7685b-6a9d-47fd-abeb-c3dc31f35801 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.413233] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662519, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.433990] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd28b6e2-c1b3-47e7-a64e-56b18949d6fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.444835] env[65788]: INFO nova.compute.manager [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Took 42.10 seconds to build instance. [ 878.446271] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662518, 'name': CreateVM_Task, 'duration_secs': 0.87129} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.447087] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.448705] env[65788]: WARNING neutronclient.v2_0.client [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 878.448705] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.448705] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.449024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 878.449313] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-132e1cb5-d62d-4da7-8e2c-cb6daac20ba4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.487205] env[65788]: DEBUG nova.network.neutron [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Updating instance_info_cache with network_info: [{"id": "421c2150-8fa1-4607-a700-cba48466f00a", "address": "fa:16:3e:77:81:88", "network": {"id": "34197f1e-7d17-4dd1-bbbb-e2afd31e35b3", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2000675529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a93220fe73e941a58f8df639e3764fd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap421c2150-8f", "ovs_interfaceid": "421c2150-8fa1-4607-a700-cba48466f00a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 878.490357] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd11651d-69dd-4c52-8599-5eca9b9e5825 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.495854] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 878.495854] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5211fcab-ff6c-608c-e2ef-9e449691eae7" [ 878.495854] env[65788]: _type = "Task" [ 878.495854] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.505663] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d81ce9-3d95-4f5c-be82-113a75a4c743 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.514804] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5211fcab-ff6c-608c-e2ef-9e449691eae7, 'name': SearchDatastore_Task, 'duration_secs': 0.012899} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.515583] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.515825] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.516079] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.516281] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 878.516524] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.517162] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48e1f607-24d2-4044-8f37-fac460141610 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.528264] env[65788]: DEBUG nova.compute.provider_tree [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.539434] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.539640] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.541602] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f8a6338-84b7-44fa-89ba-f332be53e3b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.549410] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 878.549410] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a2da4-0a9d-41eb-48a3-12cd56fa6635" [ 878.549410] env[65788]: _type = "Task" [ 878.549410] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.559987] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a2da4-0a9d-41eb-48a3-12cd56fa6635, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.732407] env[65788]: DEBUG nova.compute.manager [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 878.732711] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.734756] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381635ae-06e1-44a7-9001-eb2d1603424c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.745190] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.745190] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0e9ce7a-c0cb-440d-bd85-4bc951b23732 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.756242] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662514, 'name': CloneVM_Task, 'duration_secs': 2.214555} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.757812] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Created linked-clone VM from snapshot [ 878.758833] env[65788]: DEBUG oslo_vmware.api [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 878.758833] env[65788]: value = "task-4662520" [ 878.758833] env[65788]: _type = "Task" [ 878.758833] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.758833] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b492a692-f039-46ed-bf7f-e598cb7bd6ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.770641] env[65788]: DEBUG oslo_vmware.api [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662520, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.774116] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Uploading image 9d4ee02b-52b9-46e0-8b8e-158a076cef04 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 878.789857] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 878.790351] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-258c1b37-a79c-4f00-a9dd-65abcfd2d911 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.799671] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 878.799671] env[65788]: value = "task-4662521" [ 878.799671] env[65788]: _type = "Task" [ 878.799671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.810223] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662521, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.877192] env[65788]: DEBUG oslo_concurrency.lockutils [req-1de86288-c19c-4f11-a22d-8f03e67163c3 req-c25dcc67-f9b6-41c3-8852-3dfd19613537 service nova] Releasing lock "refresh_cache-8d7a33e6-08fb-481e-8fac-fdf13b19aa21" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.896015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "374564df-70fa-4b89-8f38-e559245e5ebf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.896264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "374564df-70fa-4b89-8f38-e559245e5ebf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.904084] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662519, 'name': Rename_Task, 'duration_secs': 0.171557} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.904832] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.904832] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d4329b5-16e8-4142-9a88-2d379521061c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.914212] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 878.914212] env[65788]: value = "task-4662522" [ 878.914212] env[65788]: _type = "Task" [ 878.914212] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.924494] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.947627] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1af51c8f-4912-4ab3-a1d5-62d0829f4f1f tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.615s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.991060] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Releasing lock "refresh_cache-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.991325] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Instance network_info: |[{"id": "421c2150-8fa1-4607-a700-cba48466f00a", "address": "fa:16:3e:77:81:88", "network": {"id": "34197f1e-7d17-4dd1-bbbb-e2afd31e35b3", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2000675529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a93220fe73e941a58f8df639e3764fd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap421c2150-8f", "ovs_interfaceid": "421c2150-8fa1-4607-a700-cba48466f00a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 878.992106] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:81:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9bb629cd-6d0f-4bed-965c-bd04a2f3ec49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '421c2150-8fa1-4607-a700-cba48466f00a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.003274] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Creating folder: Project (a93220fe73e941a58f8df639e3764fd7). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.003629] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b125bca-81d1-4e9d-b9df-d8f9a3a90c69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.018904] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Created folder: Project (a93220fe73e941a58f8df639e3764fd7) in parent group-v910111. [ 879.019253] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Creating folder: Instances. Parent ref: group-v910306. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.019545] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05de430c-c88b-4cc5-b933-01e0a224f43e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.033817] env[65788]: DEBUG nova.scheduler.client.report [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 879.037307] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Created folder: Instances in parent group-v910306. [ 879.038200] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 879.038200] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.038200] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8413e9f-b6cb-4949-a1e9-a8446a3aa392 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.065079] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a2da4-0a9d-41eb-48a3-12cd56fa6635, 'name': SearchDatastore_Task, 'duration_secs': 0.015587} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.067055] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.067055] env[65788]: value = "task-4662525" [ 879.067055] env[65788]: _type = "Task" [ 879.067055] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.067314] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0311683c-d5b1-437d-bffc-bb6d14466e41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.077187] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 879.077187] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52176505-d892-115c-6cb5-967556bbfa8f" [ 879.077187] env[65788]: _type = "Task" [ 879.077187] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.080759] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662525, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.090582] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52176505-d892-115c-6cb5-967556bbfa8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.277102] env[65788]: DEBUG oslo_vmware.api [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662520, 'name': PowerOffVM_Task, 'duration_secs': 0.316326} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.277571] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.277838] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.278272] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07eeb6ef-144c-4567-8d99-32b1e7ce0cb7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.311581] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662521, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.363650] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.363893] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.364163] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Deleting the datastore file [datastore1] 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.364501] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c7dafcd-c6df-457b-bf0b-ab279f6da77f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.376281] env[65788]: DEBUG oslo_vmware.api [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for the task: (returnval){ [ 879.376281] env[65788]: value = "task-4662527" [ 879.376281] env[65788]: _type = "Task" [ 879.376281] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.388834] env[65788]: DEBUG oslo_vmware.api [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662527, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.398741] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 879.429812] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662522, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.539513] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.285s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.542397] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.384s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.544542] env[65788]: INFO nova.compute.claims [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.568680] env[65788]: INFO nova.scheduler.client.report [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted allocations for instance eceafff8-7d28-4b9b-ade6-5736d5977efa [ 879.584236] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662525, 'name': CreateVM_Task, 'duration_secs': 0.491506} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.587453] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.588067] env[65788]: WARNING neutronclient.v2_0.client [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.588445] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.588649] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.588992] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 879.589628] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ada11616-5af8-4102-8aa0-b74627e77a5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.597419] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52176505-d892-115c-6cb5-967556bbfa8f, 'name': SearchDatastore_Task, 'duration_secs': 0.014096} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.598911] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.599212] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 8d7a33e6-08fb-481e-8fac-fdf13b19aa21/8d7a33e6-08fb-481e-8fac-fdf13b19aa21.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.599559] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 879.599559] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52014542-7779-c6c3-7082-2cf0604bd183" [ 879.599559] env[65788]: _type = "Task" [ 879.599559] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.600084] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6f7a9f1-64ce-46e3-b84f-d80f38cb252f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.612091] env[65788]: DEBUG nova.compute.manager [req-b448b557-5027-4eaa-847f-2da10b4f1fc6 req-368f58bf-37c5-46fb-afac-604ec336c55f service nova] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Received event network-vif-deleted-4a2afbad-2811-48fa-aee9-aeefd31b60d1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 879.618859] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 879.618859] env[65788]: value = "task-4662528" [ 879.618859] env[65788]: _type = "Task" [ 879.618859] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.619392] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52014542-7779-c6c3-7082-2cf0604bd183, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.630169] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.815041] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662521, 'name': Destroy_Task, 'duration_secs': 0.552007} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.815041] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Destroyed the VM [ 879.815041] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 879.815041] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-adbf4d2f-f7a0-45fd-84ef-2192053d85e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.822782] env[65788]: DEBUG nova.compute.manager [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Received event network-changed-421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 879.823059] env[65788]: DEBUG nova.compute.manager [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Refreshing instance network info cache due to event network-changed-421c2150-8fa1-4607-a700-cba48466f00a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 879.823321] env[65788]: DEBUG oslo_concurrency.lockutils [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Acquiring lock "refresh_cache-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.823525] env[65788]: DEBUG oslo_concurrency.lockutils [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Acquired lock "refresh_cache-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.823675] env[65788]: DEBUG nova.network.neutron [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Refreshing network info cache for port 421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 879.827691] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 879.827691] env[65788]: value = "task-4662529" [ 879.827691] env[65788]: _type = "Task" [ 879.827691] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.840414] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662529, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.888781] env[65788]: DEBUG oslo_vmware.api [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Task: {'id': task-4662527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.313433} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.888781] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 879.888781] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 879.888781] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 879.888781] env[65788]: INFO nova.compute.manager [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Took 1.16 seconds to destroy the instance on the hypervisor. [ 879.889111] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 879.889339] env[65788]: DEBUG nova.compute.manager [-] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 879.889437] env[65788]: DEBUG nova.network.neutron [-] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 879.889690] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.890412] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 879.890687] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 879.910111] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.911028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.925379] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.929981] env[65788]: DEBUG oslo_vmware.api [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662522, 'name': PowerOnVM_Task, 'duration_secs': 0.524806} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.930126] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.930318] env[65788]: INFO nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Took 8.75 seconds to spawn the instance on the hypervisor. [ 879.930814] env[65788]: DEBUG nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 879.931894] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 879.934387] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9421bdb9-fbea-4b3b-a918-2c8a2fd44084 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.076890] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9a2436e7-c7e7-4c55-9bae-a63c3039650a tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "eceafff8-7d28-4b9b-ade6-5736d5977efa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.081s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.115561] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52014542-7779-c6c3-7082-2cf0604bd183, 'name': SearchDatastore_Task, 'duration_secs': 0.020785} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.115909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.116172] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.116456] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.116612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 880.116818] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.117578] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-133cb445-6565-4081-95f8-2e62bd9f77d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.134103] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662528, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.138443] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.138657] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.139558] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9619fc3e-ef90-46ba-807b-906fdaf30194 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.147070] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 880.147070] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a3cf8b-ae0a-715d-5cc0-541a280b379d" [ 880.147070] env[65788]: _type = "Task" [ 880.147070] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.163059] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a3cf8b-ae0a-715d-5cc0-541a280b379d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.328901] env[65788]: WARNING neutronclient.v2_0.client [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.329800] env[65788]: WARNING openstack [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.330389] env[65788]: WARNING openstack [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.347329] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662529, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.414490] env[65788]: DEBUG nova.compute.utils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 880.436697] env[65788]: WARNING openstack [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.436697] env[65788]: WARNING openstack [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.454111] env[65788]: INFO nova.compute.manager [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Took 43.78 seconds to build instance. [ 880.506899] env[65788]: WARNING neutronclient.v2_0.client [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 880.506899] env[65788]: WARNING openstack [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 880.506899] env[65788]: WARNING openstack [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 880.600755] env[65788]: DEBUG nova.network.neutron [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Updated VIF entry in instance network info cache for port 421c2150-8fa1-4607-a700-cba48466f00a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 880.601181] env[65788]: DEBUG nova.network.neutron [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Updating instance_info_cache with network_info: [{"id": "421c2150-8fa1-4607-a700-cba48466f00a", "address": "fa:16:3e:77:81:88", "network": {"id": "34197f1e-7d17-4dd1-bbbb-e2afd31e35b3", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2000675529-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a93220fe73e941a58f8df639e3764fd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9bb629cd-6d0f-4bed-965c-bd04a2f3ec49", "external-id": "nsx-vlan-transportzone-848", "segmentation_id": 848, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap421c2150-8f", "ovs_interfaceid": "421c2150-8fa1-4607-a700-cba48466f00a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 880.633860] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637207} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.634645] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 8d7a33e6-08fb-481e-8fac-fdf13b19aa21/8d7a33e6-08fb-481e-8fac-fdf13b19aa21.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.634645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.635672] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-584197fa-b8ce-407a-8603-916e050c3673 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.646528] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 880.646528] env[65788]: value = "task-4662530" [ 880.646528] env[65788]: _type = "Task" [ 880.646528] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.663340] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662530, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.667152] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a3cf8b-ae0a-715d-5cc0-541a280b379d, 'name': SearchDatastore_Task, 'duration_secs': 0.056749} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.668038] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37eb1e65-b6bb-4b68-859a-13bfd0569659 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.675506] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 880.675506] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525b1ee1-ea2b-aa91-34d3-2fe2e480ebd1" [ 880.675506] env[65788]: _type = "Task" [ 880.675506] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.684637] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525b1ee1-ea2b-aa91-34d3-2fe2e480ebd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.746403] env[65788]: DEBUG nova.network.neutron [-] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 880.846687] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662529, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.866042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "f75000e6-f3bf-4951-a541-3aa185339054" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.866319] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.866626] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "f75000e6-f3bf-4951-a541-3aa185339054-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.866806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.866961] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.869320] env[65788]: INFO nova.compute.manager [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Terminating instance [ 880.919244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.958071] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c5c68fdd-8e03-4814-ae56-04784b735f77 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.289s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.985105] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953995f9-eb9c-4211-9394-106783254c70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.998893] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1402a3fa-dc97-4c05-b75c-089c4fafea65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.037554] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b888dcaf-187b-4510-a722-d8fbf85790ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.046248] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f368ae8-8221-4a19-b89e-4322b476b7b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.061109] env[65788]: DEBUG nova.compute.provider_tree [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.104606] env[65788]: DEBUG oslo_concurrency.lockutils [req-357fcdda-b126-497f-ab70-03b319788107 req-a827c9b7-5ff2-4718-92ca-68386d1b445c service nova] Releasing lock "refresh_cache-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.158048] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662530, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074472} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.158353] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.159159] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1baed3e9-89a4-4b82-a450-55040ee35aa9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.186921] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 8d7a33e6-08fb-481e-8fac-fdf13b19aa21/8d7a33e6-08fb-481e-8fac-fdf13b19aa21.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.189923] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4eead724-6b97-4190-9676-50f361e43b98 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.211872] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525b1ee1-ea2b-aa91-34d3-2fe2e480ebd1, 'name': SearchDatastore_Task, 'duration_secs': 0.010781} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.213596] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.214126] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468/34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.214236] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 881.214236] env[65788]: value = "task-4662531" [ 881.214236] env[65788]: _type = "Task" [ 881.214236] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.214446] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-667abf5b-745e-450f-b3cc-f5ed5c2f1f02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.229214] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662531, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.230999] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 881.230999] env[65788]: value = "task-4662532" [ 881.230999] env[65788]: _type = "Task" [ 881.230999] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.249792] env[65788]: INFO nova.compute.manager [-] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Took 1.36 seconds to deallocate network for instance. [ 881.349518] env[65788]: DEBUG oslo_vmware.api [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662529, 'name': RemoveSnapshot_Task, 'duration_secs': 1.283536} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.349863] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 881.375916] env[65788]: DEBUG nova.compute.manager [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 881.376241] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.377254] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109c5b66-0d3c-42ff-8761-b1c8d9eea9b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.386934] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.388070] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07429dd8-f5e2-4a7e-8736-cba2afde3ded {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.396814] env[65788]: DEBUG oslo_vmware.api [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 881.396814] env[65788]: value = "task-4662533" [ 881.396814] env[65788]: _type = "Task" [ 881.396814] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.407705] env[65788]: DEBUG oslo_vmware.api [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.564252] env[65788]: DEBUG nova.scheduler.client.report [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 881.677199] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.677536] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 881.730366] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662531, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.741473] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662532, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.756595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.856015] env[65788]: WARNING nova.compute.manager [None req-496987d7-e923-4279-851a-9c20fc06f709 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Image not found during snapshot: nova.exception.ImageNotFound: Image 9d4ee02b-52b9-46e0-8b8e-158a076cef04 could not be found. [ 881.908808] env[65788]: DEBUG oslo_vmware.api [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662533, 'name': PowerOffVM_Task, 'duration_secs': 0.425766} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.909128] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.909471] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.909566] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e6a4cda-6b2a-4da4-b129-05c9c623741c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.988054] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.988382] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.988579] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore1] f75000e6-f3bf-4951-a541-3aa185339054 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.988883] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9595eec-4e32-40e0-af0f-f28267774e12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.999640] env[65788]: DEBUG oslo_vmware.api [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 881.999640] env[65788]: value = "task-4662535" [ 881.999640] env[65788]: _type = "Task" [ 881.999640] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.005205] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.005320] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.005524] env[65788]: INFO nova.compute.manager [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Attaching volume 0d9578e1-90ca-4aaa-8fac-028b7f80c67a to /dev/sdb [ 882.015636] env[65788]: DEBUG oslo_vmware.api [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662535, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.068785] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b529844e-d3fb-4c81-aab4-c598b9d5a6b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.074179] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.075297] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 882.081950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.264s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.084342] env[65788]: INFO nova.compute.claims [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.095349] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1267d357-877d-430e-aeac-758051056d12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.115806] env[65788]: DEBUG nova.virt.block_device [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updating existing volume attachment record: 9d4c7a73-63e3-4e8e-9f53-545f769f9100 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 882.155503] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.157309] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.181554] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 882.231177] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662531, 'name': ReconfigVM_Task, 'duration_secs': 0.645034} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.231177] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 8d7a33e6-08fb-481e-8fac-fdf13b19aa21/8d7a33e6-08fb-481e-8fac-fdf13b19aa21.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.231177] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cc9fa24-0ce7-4333-b606-d49b2b4fbd05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.244651] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662532, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562519} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.246739] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468/34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.247088] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.247796] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 882.247796] env[65788]: value = "task-4662536" [ 882.247796] env[65788]: _type = "Task" [ 882.247796] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.247998] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c423963-2b99-461f-a086-6798a62835a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.264949] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 882.264949] env[65788]: value = "task-4662537" [ 882.264949] env[65788]: _type = "Task" [ 882.264949] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.276868] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662537, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.491026] env[65788]: DEBUG nova.compute.manager [req-afaaf441-e056-494a-b71a-eaca6283fbd5 req-5ab37930-4821-4cab-8397-4980e8d82c68 service nova] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Received event network-vif-deleted-6fe0678d-9d42-48ef-ace8-37c30ef0cc23 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 882.510161] env[65788]: DEBUG oslo_vmware.api [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662535, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154976} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.510423] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.510601] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.510773] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.510940] env[65788]: INFO nova.compute.manager [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Took 1.13 seconds to destroy the instance on the hypervisor. [ 882.511353] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 882.511536] env[65788]: DEBUG nova.compute.manager [-] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 882.511646] env[65788]: DEBUG nova.network.neutron [-] [instance: f75000e6-f3bf-4951-a541-3aa185339054] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 882.511902] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.512689] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.512946] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.585523] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.593193] env[65788]: DEBUG nova.compute.utils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 882.596040] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 882.596040] env[65788]: DEBUG nova.network.neutron [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 882.596177] env[65788]: WARNING neutronclient.v2_0.client [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.597035] env[65788]: WARNING neutronclient.v2_0.client [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 882.597937] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 882.598471] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 882.645454] env[65788]: DEBUG nova.policy [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e879429f0504963a38563a287f114f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '021bf6dd87e340969a57e7e5b33268e7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 882.659462] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 882.714021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.764084] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662536, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.779297] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662537, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133759} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.779297] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.779297] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77bd616-fd51-4c70-b3be-82c39ec9eebd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.806147] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468/34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.808753] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f611e15e-3c84-42ad-886f-165b47397552 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.833933] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 882.833933] env[65788]: value = "task-4662541" [ 882.833933] env[65788]: _type = "Task" [ 882.833933] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.844379] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662541, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.936090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.936090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.936090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.936387] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.936567] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.939102] env[65788]: INFO nova.compute.manager [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Terminating instance [ 883.072478] env[65788]: DEBUG nova.network.neutron [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Successfully created port: 4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 883.106993] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 883.182887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 883.266502] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662536, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.326653] env[65788]: DEBUG nova.network.neutron [-] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 883.348536] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.445383] env[65788]: DEBUG nova.compute.manager [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 883.445383] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.449985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132d0b3d-4e21-4880-a771-371d7835cf5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.460766] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.460875] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fb739b1-0181-479c-b6de-197d9d3ead90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.470712] env[65788]: DEBUG oslo_vmware.api [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 883.470712] env[65788]: value = "task-4662542" [ 883.470712] env[65788]: _type = "Task" [ 883.470712] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.487686] env[65788]: DEBUG oslo_vmware.api [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.594445] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45dc8e8-48e8-46f2-a4c4-460f6ec0befa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.605226] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f58be4-335e-4626-8cac-d3de03a5c592 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.656139] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eb5325-df30-4410-857e-2265e610fe9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.669975] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782749dd-3882-4735-a71c-93182e21b696 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.685451] env[65788]: DEBUG nova.compute.provider_tree [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.768350] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662536, 'name': Rename_Task, 'duration_secs': 1.243328} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.768599] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.768787] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-402978fe-901b-44e1-b9ab-955153529373 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.776371] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 883.776371] env[65788]: value = "task-4662543" [ 883.776371] env[65788]: _type = "Task" [ 883.776371] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.786878] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.832683] env[65788]: INFO nova.compute.manager [-] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Took 1.32 seconds to deallocate network for instance. [ 883.847790] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.985493] env[65788]: DEBUG oslo_vmware.api [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662542, 'name': PowerOffVM_Task, 'duration_secs': 0.203465} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.985804] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.985961] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 883.986304] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ce82b8d-1229-4e0b-899a-992b1195c52a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.119391] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 884.188676] env[65788]: DEBUG nova.scheduler.client.report [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 884.227052] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 884.227353] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 884.227512] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 884.227764] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 884.227929] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 884.228101] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 884.228324] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.228497] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 884.228678] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 884.228844] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 884.229058] env[65788]: DEBUG nova.virt.hardware [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 884.230400] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc293d6-3afc-4a35-a8e1-d3a18b09fb89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.240590] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f179cd99-1b96-4fc9-9fca-04f6bd716b52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.271959] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.272321] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.272725] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleting the datastore file [datastore1] d042a386-0d16-4e2d-a8e7-7641c07bcb35 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.272827] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d40ec6f7-4a82-4dcd-8452-01d18288b226 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.283192] env[65788]: DEBUG oslo_vmware.api [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 884.283192] env[65788]: value = "task-4662545" [ 884.283192] env[65788]: _type = "Task" [ 884.283192] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.292497] env[65788]: DEBUG oslo_vmware.api [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662543, 'name': PowerOnVM_Task, 'duration_secs': 0.48451} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.293196] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.294022] env[65788]: INFO nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Took 10.96 seconds to spawn the instance on the hypervisor. [ 884.294022] env[65788]: DEBUG nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 884.294400] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e771b52-815e-453a-ad47-37fcb578a7b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.300406] env[65788]: DEBUG oslo_vmware.api [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662545, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.343576] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.346925] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662541, 'name': ReconfigVM_Task, 'duration_secs': 1.129962} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.347638] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468/34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.349325] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-884ecc42-d135-4261-a337-4ca75b01fcef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.360038] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 884.360038] env[65788]: value = "task-4662546" [ 884.360038] env[65788]: _type = "Task" [ 884.360038] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.369236] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662546, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.549583] env[65788]: DEBUG nova.compute.manager [req-871fab16-f97b-4336-b133-c21cce380f4d req-47e561af-a858-4f34-9d32-704eb5d02c5c service nova] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Received event network-vif-deleted-9e5e0e29-c770-41e3-9766-2be8e6f3dc96 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 884.694402] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.694936] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 884.697841] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.821s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.699404] env[65788]: INFO nova.compute.claims [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.725146] env[65788]: DEBUG nova.network.neutron [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Successfully updated port: 4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 884.795012] env[65788]: DEBUG oslo_vmware.api [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662545, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223195} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.795349] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.795475] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.795644] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.795811] env[65788]: INFO nova.compute.manager [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Took 1.35 seconds to destroy the instance on the hypervisor. [ 884.796079] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 884.796297] env[65788]: DEBUG nova.compute.manager [-] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 884.796395] env[65788]: DEBUG nova.network.neutron [-] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 884.796646] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.797200] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 884.797453] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 884.817410] env[65788]: INFO nova.compute.manager [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Took 47.63 seconds to build instance. [ 884.866578] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 884.872814] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662546, 'name': Rename_Task, 'duration_secs': 0.185338} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.873125] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.873417] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4ea42fd-d07f-4c04-a77e-417d4577e256 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.881511] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 884.881511] env[65788]: value = "task-4662548" [ 884.881511] env[65788]: _type = "Task" [ 884.881511] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.890615] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.207027] env[65788]: DEBUG nova.compute.utils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 885.210652] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 885.211386] env[65788]: DEBUG nova.network.neutron [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 885.212158] env[65788]: WARNING neutronclient.v2_0.client [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.212698] env[65788]: WARNING neutronclient.v2_0.client [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.213520] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.214205] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.231580] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.231580] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.231580] env[65788]: DEBUG nova.network.neutron [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 885.265358] env[65788]: DEBUG nova.compute.manager [req-2fd024a7-b9da-4519-baab-6b495bf0eb54 req-4f1ea566-2a6d-41eb-9366-58576b0e32d7 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Received event network-vif-deleted-48d24f13-8cff-4392-b381-04ec0ce66f8d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 885.265550] env[65788]: INFO nova.compute.manager [req-2fd024a7-b9da-4519-baab-6b495bf0eb54 req-4f1ea566-2a6d-41eb-9366-58576b0e32d7 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Neutron deleted interface 48d24f13-8cff-4392-b381-04ec0ce66f8d; detaching it from the instance and deleting it from the info cache [ 885.265720] env[65788]: DEBUG nova.network.neutron [req-2fd024a7-b9da-4519-baab-6b495bf0eb54 req-4f1ea566-2a6d-41eb-9366-58576b0e32d7 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.305140] env[65788]: DEBUG nova.policy [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd9a17b2eb1434b0e9297442bfc28d152', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e7aa18dcc41403389518d700c2c8b8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 885.319325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1576d4c5-554d-4adf-9d99-e0d80407c984 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.145s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.397081] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662548, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.637571] env[65788]: DEBUG nova.network.neutron [-] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 885.648904] env[65788]: DEBUG nova.network.neutron [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Successfully created port: d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 885.712342] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 885.737859] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.737859] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.770160] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-deeba0b7-8fd5-4cb1-a104-dc2ae7ef5d3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.781782] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f53df2-a190-498d-8d6d-9abbfe6318f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.830307] env[65788]: DEBUG nova.compute.manager [req-2fd024a7-b9da-4519-baab-6b495bf0eb54 req-4f1ea566-2a6d-41eb-9366-58576b0e32d7 service nova] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Detach interface failed, port_id=48d24f13-8cff-4392-b381-04ec0ce66f8d, reason: Instance d042a386-0d16-4e2d-a8e7-7641c07bcb35 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 885.831618] env[65788]: DEBUG nova.network.neutron [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 885.868111] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.868598] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.894134] env[65788]: DEBUG oslo_vmware.api [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662548, 'name': PowerOnVM_Task, 'duration_secs': 0.741935} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.897119] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.897357] env[65788]: INFO nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Took 9.41 seconds to spawn the instance on the hypervisor. [ 885.897724] env[65788]: DEBUG nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 885.899447] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b4fc5f-d4bb-420a-b83c-e0d6c5d675a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.960594] env[65788]: WARNING neutronclient.v2_0.client [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 885.961270] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.961633] env[65788]: WARNING openstack [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.128157] env[65788]: DEBUG nova.network.neutron [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [{"id": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "address": "fa:16:3e:88:8c:66", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ac3e7fa-b2", "ovs_interfaceid": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.141385] env[65788]: INFO nova.compute.manager [-] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Took 1.34 seconds to deallocate network for instance. [ 886.276884] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b83a01b-5b82-413b-a011-4a6e246c1762 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.287185] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839464f8-bd2e-4c76-b5d4-e011d58c855e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.345086] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4f5a19-d848-447e-a67d-224cf3ba0c6a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.354662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6382c5c4-2613-48ce-9aee-779336adca11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.370555] env[65788]: DEBUG nova.compute.provider_tree [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.420554] env[65788]: INFO nova.compute.manager [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Took 43.95 seconds to build instance. [ 886.588636] env[65788]: DEBUG nova.compute.manager [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Received event network-vif-plugged-4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 886.588884] env[65788]: DEBUG oslo_concurrency.lockutils [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 886.589168] env[65788]: DEBUG oslo_concurrency.lockutils [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 886.589355] env[65788]: DEBUG oslo_concurrency.lockutils [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.589525] env[65788]: DEBUG nova.compute.manager [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] No waiting events found dispatching network-vif-plugged-4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 886.589694] env[65788]: WARNING nova.compute.manager [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Received unexpected event network-vif-plugged-4ac3e7fa-b26e-4239-be70-ae1397fbba52 for instance with vm_state building and task_state spawning. [ 886.589868] env[65788]: DEBUG nova.compute.manager [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Received event network-changed-4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 886.590105] env[65788]: DEBUG nova.compute.manager [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Refreshing instance network info cache due to event network-changed-4ac3e7fa-b26e-4239-be70-ae1397fbba52. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 886.590517] env[65788]: DEBUG oslo_concurrency.lockutils [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Acquiring lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.632186] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.632595] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Instance network_info: |[{"id": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "address": "fa:16:3e:88:8c:66", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ac3e7fa-b2", "ovs_interfaceid": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 886.632962] env[65788]: DEBUG oslo_concurrency.lockutils [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Acquired lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.633174] env[65788]: DEBUG nova.network.neutron [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Refreshing network info cache for port 4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 886.638034] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:8c:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '706c9762-1cf8-4770-897d-377d0d927773', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ac3e7fa-b26e-4239-be70-ae1397fbba52', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.642869] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Creating folder: Project (021bf6dd87e340969a57e7e5b33268e7). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.643524] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3427f27-b2a8-4e2f-b307-ae40aaa14ed4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.648661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 886.657763] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Created folder: Project (021bf6dd87e340969a57e7e5b33268e7) in parent group-v910111. [ 886.657947] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Creating folder: Instances. Parent ref: group-v910311. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.658212] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cbef9d5-2896-481f-8649-d21a21194595 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.672658] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Created folder: Instances in parent group-v910311. [ 886.672962] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 886.673317] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.673564] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56db4492-1368-4fee-858b-361a209255be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.692499] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 886.692813] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 886.694070] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dd5c83-c51f-4ef6-bd43-edf5090d8695 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.714471] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d20dd1-2e13-487c-a9e1-8b5f672260bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.717937] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.717937] env[65788]: value = "task-4662551" [ 886.717937] env[65788]: _type = "Task" [ 886.717937] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.736873] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 886.747135] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a/volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.748273] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2896559-6c37-4e5f-aff8-9a61639fa8e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.768846] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662551, 'name': CreateVM_Task} progress is 15%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.778158] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 886.778447] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 886.778604] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 886.778789] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 886.778934] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 886.779095] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 886.779309] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.779467] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 886.779634] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 886.779818] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 886.780056] env[65788]: DEBUG nova.virt.hardware [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 886.781211] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb3719e-d3b6-4221-9e61-258045ef272f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.787655] env[65788]: DEBUG oslo_vmware.api [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 886.787655] env[65788]: value = "task-4662552" [ 886.787655] env[65788]: _type = "Task" [ 886.787655] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.795879] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3d8e26-e315-4edc-85d9-60d6f9d6ace6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.803660] env[65788]: DEBUG oslo_vmware.api [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662552, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.874533] env[65788]: DEBUG nova.scheduler.client.report [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 886.922438] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d724b288-4763-411c-ab81-b200fd6ad454 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.460s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.145445] env[65788]: WARNING neutronclient.v2_0.client [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.146068] env[65788]: WARNING openstack [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.146477] env[65788]: WARNING openstack [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.235629] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662551, 'name': CreateVM_Task, 'duration_secs': 0.435107} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.235912] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.237041] env[65788]: WARNING neutronclient.v2_0.client [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.237488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.237736] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.238359] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 887.238586] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15aa5e9a-da33-4cce-9e1e-5204f35de97b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.246850] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 887.246850] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529cbed4-f2c1-53cc-ef89-143776f7b6d8" [ 887.246850] env[65788]: _type = "Task" [ 887.246850] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.266909] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529cbed4-f2c1-53cc-ef89-143776f7b6d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.283359] env[65788]: DEBUG nova.network.neutron [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Successfully updated port: d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 887.289041] env[65788]: WARNING openstack [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.289341] env[65788]: WARNING openstack [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.309473] env[65788]: DEBUG oslo_vmware.api [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662552, 'name': ReconfigVM_Task, 'duration_secs': 0.522738} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.309774] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfigured VM instance instance-00000038 to attach disk [datastore1] volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a/volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.314947] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c411aacf-298c-4537-bb96-5b249876d2cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.337796] env[65788]: DEBUG oslo_vmware.api [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 887.337796] env[65788]: value = "task-4662553" [ 887.337796] env[65788]: _type = "Task" [ 887.337796] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.348968] env[65788]: DEBUG oslo_vmware.api [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662553, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.362807] env[65788]: WARNING neutronclient.v2_0.client [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.363524] env[65788]: WARNING openstack [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.364430] env[65788]: WARNING openstack [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.383236] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.685s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.384551] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 887.386588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.930s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.388066] env[65788]: INFO nova.compute.claims [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.465427] env[65788]: DEBUG nova.network.neutron [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updated VIF entry in instance network info cache for port 4ac3e7fa-b26e-4239-be70-ae1397fbba52. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 887.465876] env[65788]: DEBUG nova.network.neutron [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [{"id": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "address": "fa:16:3e:88:8c:66", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ac3e7fa-b2", "ovs_interfaceid": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 887.627720] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "interface-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.627984] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "interface-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.628371] env[65788]: DEBUG nova.objects.instance [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lazy-loading 'flavor' on Instance uuid 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.765474] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529cbed4-f2c1-53cc-ef89-143776f7b6d8, 'name': SearchDatastore_Task, 'duration_secs': 0.019281} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.765767] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.766063] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.766258] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.766576] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.766793] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.767105] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-634d8520-d7d7-4e09-8037-ea8910c900d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.782833] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.783054] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.783913] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be5d41ff-1bff-4a4a-a72e-4b5964274f7a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.787194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "refresh_cache-5e27dbf0-44d2-46a4-87b2-c209afb9559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.787356] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "refresh_cache-5e27dbf0-44d2-46a4-87b2-c209afb9559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.787622] env[65788]: DEBUG nova.network.neutron [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 887.792693] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 887.792693] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5236ac04-7b04-b1e2-81f6-9b1c026f4f9b" [ 887.792693] env[65788]: _type = "Task" [ 887.792693] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.802112] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5236ac04-7b04-b1e2-81f6-9b1c026f4f9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.849515] env[65788]: DEBUG oslo_vmware.api [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662553, 'name': ReconfigVM_Task, 'duration_secs': 0.192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.849848] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 887.893108] env[65788]: DEBUG nova.compute.utils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 887.896639] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 887.896858] env[65788]: DEBUG nova.network.neutron [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 887.897706] env[65788]: WARNING neutronclient.v2_0.client [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.897706] env[65788]: WARNING neutronclient.v2_0.client [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 887.898041] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.899678] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.945971] env[65788]: DEBUG nova.policy [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a1e3cda6bf9439982465aec80977cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1713c7002a413fb27180469fded83e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 887.969186] env[65788]: DEBUG oslo_concurrency.lockutils [req-8764b710-9538-4634-a7f7-b2ac4bf9fc86 req-41a7ced6-750e-446b-9dfa-d3c1c5bc776e service nova] Releasing lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.134168] env[65788]: WARNING neutronclient.v2_0.client [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.135100] env[65788]: WARNING openstack [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.135554] env[65788]: WARNING openstack [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.141393] env[65788]: DEBUG nova.objects.instance [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lazy-loading 'pci_requests' on Instance uuid 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.286505] env[65788]: DEBUG nova.network.neutron [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Successfully created port: 08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 888.292160] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.292160] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.316730] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5236ac04-7b04-b1e2-81f6-9b1c026f4f9b, 'name': SearchDatastore_Task, 'duration_secs': 0.013148} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.317611] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e971b30e-cfec-4dd6-9688-bb280cfce269 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.324707] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 888.324707] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529de06b-9ce6-a79c-2835-b78a397865eb" [ 888.324707] env[65788]: _type = "Task" [ 888.324707] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.336243] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529de06b-9ce6-a79c-2835-b78a397865eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.337231] env[65788]: DEBUG nova.network.neutron [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 888.358758] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.358758] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.397956] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 888.440243] env[65788]: WARNING neutronclient.v2_0.client [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.441520] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.441934] env[65788]: WARNING openstack [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.542243] env[65788]: DEBUG nova.network.neutron [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Updating instance_info_cache with network_info: [{"id": "d6457e25-41bd-486e-bd24-5ec2f17cfbb3", "address": "fa:16:3e:0b:a0:e5", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6457e25-41", "ovs_interfaceid": "d6457e25-41bd-486e-bd24-5ec2f17cfbb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 888.644351] env[65788]: DEBUG nova.objects.base [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Object Instance<34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468> lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 888.644539] env[65788]: DEBUG nova.network.neutron [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 888.644867] env[65788]: WARNING neutronclient.v2_0.client [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.645177] env[65788]: WARNING neutronclient.v2_0.client [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 888.645772] env[65788]: WARNING openstack [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 888.646116] env[65788]: WARNING openstack [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 888.779802] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b0f5112-122d-4988-b7e5-5999350572d7 tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "interface-34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.152s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.820388] env[65788]: DEBUG nova.compute.manager [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Received event network-vif-plugged-d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 888.820596] env[65788]: DEBUG oslo_concurrency.lockutils [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Acquiring lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 888.820853] env[65788]: DEBUG oslo_concurrency.lockutils [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.820960] env[65788]: DEBUG oslo_concurrency.lockutils [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.821206] env[65788]: DEBUG nova.compute.manager [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] No waiting events found dispatching network-vif-plugged-d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 888.821672] env[65788]: WARNING nova.compute.manager [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Received unexpected event network-vif-plugged-d6457e25-41bd-486e-bd24-5ec2f17cfbb3 for instance with vm_state building and task_state spawning. [ 888.822877] env[65788]: DEBUG nova.compute.manager [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Received event network-changed-d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 888.822877] env[65788]: DEBUG nova.compute.manager [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Refreshing instance network info cache due to event network-changed-d6457e25-41bd-486e-bd24-5ec2f17cfbb3. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 888.822877] env[65788]: DEBUG oslo_concurrency.lockutils [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Acquiring lock "refresh_cache-5e27dbf0-44d2-46a4-87b2-c209afb9559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.837253] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529de06b-9ce6-a79c-2835-b78a397865eb, 'name': SearchDatastore_Task, 'duration_secs': 0.014491} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.840136] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.840317] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d8103991-fd1f-4a57-81ce-1a47dc4defe0/d8103991-fd1f-4a57-81ce-1a47dc4defe0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.840791] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ed4eeea-f186-4baf-bae3-c92470a337e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.848408] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 888.848408] env[65788]: value = "task-4662554" [ 888.848408] env[65788]: _type = "Task" [ 888.848408] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.859267] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.864797] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055672a3-ea80-4091-bad9-20803fb518ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.873016] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc802fd-c09c-4131-a4c2-5d8dee0afc76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.908773] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675e3538-9b30-418a-a762-cdb763755cad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.919640] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6720b286-6e81-4b77-b91a-2bbd2fcf7182 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.935231] env[65788]: DEBUG nova.compute.provider_tree [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.033634] env[65788]: DEBUG nova.objects.instance [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'flavor' on Instance uuid f160e4a5-1e91-495e-800e-116ec435d8e1 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.047395] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "refresh_cache-5e27dbf0-44d2-46a4-87b2-c209afb9559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.047778] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Instance network_info: |[{"id": "d6457e25-41bd-486e-bd24-5ec2f17cfbb3", "address": "fa:16:3e:0b:a0:e5", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6457e25-41", "ovs_interfaceid": "d6457e25-41bd-486e-bd24-5ec2f17cfbb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 889.048838] env[65788]: DEBUG oslo_concurrency.lockutils [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Acquired lock "refresh_cache-5e27dbf0-44d2-46a4-87b2-c209afb9559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.049035] env[65788]: DEBUG nova.network.neutron [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Refreshing network info cache for port d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 889.050300] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:a0:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6457e25-41bd-486e-bd24-5ec2f17cfbb3', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.058184] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 889.059165] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.059397] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4ca4b16-c92d-4e72-9643-d77423357799 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.079939] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.079939] env[65788]: value = "task-4662555" [ 889.079939] env[65788]: _type = "Task" [ 889.079939] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.088969] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662555, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.360220] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662554, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.415350] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 889.442281] env[65788]: DEBUG nova.scheduler.client.report [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.454260] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='2b9be12192397c4f843904dea8fec0cd',container_format='bare',created_at=2025-11-21T13:50:03Z,direct_url=,disk_format='vmdk',id=e6f1a08c-b319-4a8d-9b85-bd96892a6aab,min_disk=1,min_ram=0,name='tempest-test-snap-927867648',owner='3e1713c7002a413fb27180469fded83e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-11-21T13:50:19Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 889.454661] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 889.454877] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 889.455101] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 889.455264] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 889.455417] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 889.455679] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.455865] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 889.456058] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 889.456309] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 889.456624] env[65788]: DEBUG nova.virt.hardware [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 889.457539] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9b15ce-09c4-4da0-8428-bf73c105c19b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.468860] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392ce6f4-be5d-4601-9045-8df5cdc80315 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.539483] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e72523a0-bf0a-43fd-81de-7b3c3fa3fa74 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.534s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.559807] env[65788]: WARNING neutronclient.v2_0.client [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.560670] env[65788]: WARNING openstack [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.561183] env[65788]: WARNING openstack [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.592974] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662555, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.672027] env[65788]: WARNING openstack [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.672110] env[65788]: WARNING openstack [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.746613] env[65788]: WARNING neutronclient.v2_0.client [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 889.747378] env[65788]: WARNING openstack [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 889.747722] env[65788]: WARNING openstack [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 889.863025] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662554, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.813227} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.866454] env[65788]: DEBUG nova.network.neutron [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Updated VIF entry in instance network info cache for port d6457e25-41bd-486e-bd24-5ec2f17cfbb3. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 889.867597] env[65788]: DEBUG nova.network.neutron [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Updating instance_info_cache with network_info: [{"id": "d6457e25-41bd-486e-bd24-5ec2f17cfbb3", "address": "fa:16:3e:0b:a0:e5", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6457e25-41", "ovs_interfaceid": "d6457e25-41bd-486e-bd24-5ec2f17cfbb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 889.868422] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] d8103991-fd1f-4a57-81ce-1a47dc4defe0/d8103991-fd1f-4a57-81ce-1a47dc4defe0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.868628] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.871221] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-538b736e-fd16-446f-9d70-54d7b365e473 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.874212] env[65788]: DEBUG nova.network.neutron [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Successfully updated port: 08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 889.882774] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 889.882774] env[65788]: value = "task-4662556" [ 889.882774] env[65788]: _type = "Task" [ 889.882774] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.895170] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662556, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.949434] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.950046] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 889.953123] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.676s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.953449] env[65788]: DEBUG nova.objects.instance [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lazy-loading 'resources' on Instance uuid f4a490a2-f2b1-4eac-8c1a-a18758583c70 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.092981] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662555, 'name': CreateVM_Task, 'duration_secs': 0.844283} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.093226] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.093720] env[65788]: WARNING neutronclient.v2_0.client [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.094101] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.094362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.094627] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 890.094900] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6511aa32-fcfc-404f-bd49-c620dcb3164d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.101614] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 890.101614] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5205333b-64c6-2fcf-7a9f-154d46b5b890" [ 890.101614] env[65788]: _type = "Task" [ 890.101614] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.110397] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5205333b-64c6-2fcf-7a9f-154d46b5b890, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.126102] env[65788]: DEBUG oslo_concurrency.lockutils [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.126466] env[65788]: DEBUG oslo_concurrency.lockutils [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.126718] env[65788]: DEBUG nova.compute.manager [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 890.127832] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cecff3-c635-49d1-ad15-e9970892cb60 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.135785] env[65788]: DEBUG nova.compute.manager [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 890.136339] env[65788]: DEBUG nova.objects.instance [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lazy-loading 'flavor' on Instance uuid d2d614f8-15d9-47d6-9e66-8e1f8fcc18da {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.317635] env[65788]: DEBUG nova.compute.manager [req-bb340c48-b003-4d85-9ab1-f5fade4c3e0d req-dea5460d-3b92-4bbd-b0a6-08d2596b5b1c service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Received event network-vif-plugged-08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 890.317843] env[65788]: DEBUG oslo_concurrency.lockutils [req-bb340c48-b003-4d85-9ab1-f5fade4c3e0d req-dea5460d-3b92-4bbd-b0a6-08d2596b5b1c service nova] Acquiring lock "7c88e3ef-82d2-46ac-a350-999de8091c01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 890.318052] env[65788]: DEBUG oslo_concurrency.lockutils [req-bb340c48-b003-4d85-9ab1-f5fade4c3e0d req-dea5460d-3b92-4bbd-b0a6-08d2596b5b1c service nova] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 890.318227] env[65788]: DEBUG oslo_concurrency.lockutils [req-bb340c48-b003-4d85-9ab1-f5fade4c3e0d req-dea5460d-3b92-4bbd-b0a6-08d2596b5b1c service nova] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 890.318397] env[65788]: DEBUG nova.compute.manager [req-bb340c48-b003-4d85-9ab1-f5fade4c3e0d req-dea5460d-3b92-4bbd-b0a6-08d2596b5b1c service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] No waiting events found dispatching network-vif-plugged-08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 890.318553] env[65788]: WARNING nova.compute.manager [req-bb340c48-b003-4d85-9ab1-f5fade4c3e0d req-dea5460d-3b92-4bbd-b0a6-08d2596b5b1c service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Received unexpected event network-vif-plugged-08be336c-c7bd-448b-b11d-931b7e9788c9 for instance with vm_state building and task_state spawning. [ 890.370502] env[65788]: DEBUG oslo_concurrency.lockutils [req-3ad9f162-5da2-4710-b8bf-51bfc40d6f0e req-c34b71cd-b003-4985-b813-93e59b25784d service nova] Releasing lock "refresh_cache-5e27dbf0-44d2-46a4-87b2-c209afb9559e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.377565] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "refresh_cache-7c88e3ef-82d2-46ac-a350-999de8091c01" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.377824] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "refresh_cache-7c88e3ef-82d2-46ac-a350-999de8091c01" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.378088] env[65788]: DEBUG nova.network.neutron [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 890.395163] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662556, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100577} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.395487] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.396320] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413e42be-9bef-4fe7-9ef2-61db59a9ccdc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.420158] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] d8103991-fd1f-4a57-81ce-1a47dc4defe0/d8103991-fd1f-4a57-81ce-1a47dc4defe0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.420483] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74d1295f-6a2a-407d-a98c-093c55f10faa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.442060] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 890.442060] env[65788]: value = "task-4662557" [ 890.442060] env[65788]: _type = "Task" [ 890.442060] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.457576] env[65788]: DEBUG nova.compute.utils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 890.463390] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662557, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.464343] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 890.465319] env[65788]: DEBUG nova.network.neutron [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 890.465520] env[65788]: WARNING neutronclient.v2_0.client [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.465789] env[65788]: WARNING neutronclient.v2_0.client [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 890.466538] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.466922] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.515179] env[65788]: DEBUG nova.policy [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32b4adcc8fa42fba0dbffd4582fcf69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36859c1eb994614b2a77400f811cf50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 890.615351] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5205333b-64c6-2fcf-7a9f-154d46b5b890, 'name': SearchDatastore_Task, 'duration_secs': 0.021015} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.618265] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 890.618519] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.618757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.618900] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.619089] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.619564] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65972ffd-5530-41e5-a005-f56dc009f5a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.630409] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.630409] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.634192] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12197bf4-d8dd-457f-98af-ada327145e3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.641699] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 890.641699] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e8e9f4-1274-583c-3bb1-732fdbbcd237" [ 890.641699] env[65788]: _type = "Task" [ 890.641699] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.655542] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e8e9f4-1274-583c-3bb1-732fdbbcd237, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.810898] env[65788]: INFO nova.compute.manager [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Rebuilding instance [ 890.831875] env[65788]: DEBUG nova.network.neutron [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Successfully created port: 987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 890.865446] env[65788]: DEBUG nova.compute.manager [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 890.865446] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff66f41-18b2-498d-a962-28c1d465a437 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.886472] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.887060] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.937087] env[65788]: DEBUG nova.network.neutron [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 890.954889] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662557, 'name': ReconfigVM_Task, 'duration_secs': 0.43294} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.958319] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Reconfigured VM instance instance-00000042 to attach disk [datastore1] d8103991-fd1f-4a57-81ce-1a47dc4defe0/d8103991-fd1f-4a57-81ce-1a47dc4defe0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.959500] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bde57630-6254-4f8b-8d98-97b4f2c4789e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.965386] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 890.970921] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 890.970921] env[65788]: value = "task-4662558" [ 890.970921] env[65788]: _type = "Task" [ 890.970921] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.976018] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 890.976018] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 890.994047] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662558, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.996291] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56ef85b-64a4-4043-ae5d-880fb3233b81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.007440] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a14826-357e-40df-99c0-27cb524f6b2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.045042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.045358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.045571] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 891.045757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.045923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.048077] env[65788]: INFO nova.compute.manager [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Terminating instance [ 891.058314] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600f7394-df0e-40ae-b7e3-2ceaddfa8805 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.069348] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6d9b0a-2135-4596-ba5d-6631e2637b5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.085813] env[65788]: DEBUG nova.compute.provider_tree [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.108943] env[65788]: WARNING neutronclient.v2_0.client [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 891.109644] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 891.110017] env[65788]: WARNING openstack [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 891.147863] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.148063] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa6b5e43-1bdf-45db-8e36-2f5ae12ac2e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.157741] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e8e9f4-1274-583c-3bb1-732fdbbcd237, 'name': SearchDatastore_Task, 'duration_secs': 0.015758} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.159355] env[65788]: DEBUG oslo_vmware.api [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 891.159355] env[65788]: value = "task-4662559" [ 891.159355] env[65788]: _type = "Task" [ 891.159355] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.159580] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0418ba0e-c752-4325-bc85-7e444481e224 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.170638] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 891.170638] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b047a-d93b-a699-4521-74e98b9ffbdc" [ 891.170638] env[65788]: _type = "Task" [ 891.170638] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.173843] env[65788]: DEBUG oslo_vmware.api [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662559, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.182924] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b047a-d93b-a699-4521-74e98b9ffbdc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.284015] env[65788]: DEBUG nova.network.neutron [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Updating instance_info_cache with network_info: [{"id": "08be336c-c7bd-448b-b11d-931b7e9788c9", "address": "fa:16:3e:0b:50:56", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08be336c-c7", "ovs_interfaceid": "08be336c-c7bd-448b-b11d-931b7e9788c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 891.490486] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662558, 'name': Rename_Task, 'duration_secs': 0.199083} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.490665] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.490901] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-925fc758-64ec-4c6a-9a14-3b51a475a20b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.498451] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 891.498451] env[65788]: value = "task-4662560" [ 891.498451] env[65788]: _type = "Task" [ 891.498451] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.507596] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662560, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.552369] env[65788]: DEBUG nova.compute.manager [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 891.553035] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.553769] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429ae427-fa3b-41fc-9326-0d43422fe65b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.562679] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.562679] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad8f0c04-1b03-4acd-8e9b-d9981550fbfa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.572953] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 891.572953] env[65788]: value = "task-4662561" [ 891.572953] env[65788]: _type = "Task" [ 891.572953] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.585348] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662561, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.589650] env[65788]: DEBUG nova.scheduler.client.report [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.679519] env[65788]: DEBUG oslo_vmware.api [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662559, 'name': PowerOffVM_Task, 'duration_secs': 0.232351} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.685703] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.685931] env[65788]: DEBUG nova.compute.manager [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 891.690316] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732068e8-4c8a-4661-8cee-e080e99498aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.700492] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b047a-d93b-a699-4521-74e98b9ffbdc, 'name': SearchDatastore_Task, 'duration_secs': 0.034983} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.702444] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.702824] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 5e27dbf0-44d2-46a4-87b2-c209afb9559e/5e27dbf0-44d2-46a4-87b2-c209afb9559e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.705523] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eef7394a-d034-4dd7-a2fc-91254ac692a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.714584] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 891.714584] env[65788]: value = "task-4662562" [ 891.714584] env[65788]: _type = "Task" [ 891.714584] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.723940] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.787037] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "refresh_cache-7c88e3ef-82d2-46ac-a350-999de8091c01" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.787498] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Instance network_info: |[{"id": "08be336c-c7bd-448b-b11d-931b7e9788c9", "address": "fa:16:3e:0b:50:56", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08be336c-c7", "ovs_interfaceid": "08be336c-c7bd-448b-b11d-931b7e9788c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 891.788030] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:50:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08be336c-c7bd-448b-b11d-931b7e9788c9', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.799241] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 891.799241] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.799241] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58fc32ca-9c65-4fd8-9312-5b92f1308b35 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.821279] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.821279] env[65788]: value = "task-4662563" [ 891.821279] env[65788]: _type = "Task" [ 891.821279] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.831038] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662563, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.880437] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.881205] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41cec5cd-3a1f-4be7-90eb-561ac2e8702c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.890781] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 891.890781] env[65788]: value = "task-4662564" [ 891.890781] env[65788]: _type = "Task" [ 891.890781] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.903201] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.977982] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 892.015212] env[65788]: DEBUG oslo_vmware.api [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662560, 'name': PowerOnVM_Task, 'duration_secs': 0.487582} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.015212] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.015212] env[65788]: INFO nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Took 7.90 seconds to spawn the instance on the hypervisor. [ 892.015212] env[65788]: DEBUG nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 892.016705] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53554682-ab80-4f3a-a207-4095fae887eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.022621] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 892.022869] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 892.023108] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 892.023259] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 892.024114] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 892.024114] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 892.024114] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 892.024114] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 892.024114] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 892.024358] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 892.024633] env[65788]: DEBUG nova.virt.hardware [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 892.025651] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cd843a-e927-4071-b2ea-b5097fdcaf58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.040144] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942776ee-01a8-4974-9bd0-53ad8d7b5ea6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.086439] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662561, 'name': PowerOffVM_Task, 'duration_secs': 0.275333} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.086580] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 892.086795] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 892.087091] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b022fa7-ddb4-4eda-afbf-232e880110f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.095646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.143s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.098803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.025s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 892.099292] env[65788]: DEBUG nova.objects.instance [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lazy-loading 'resources' on Instance uuid 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.133772] env[65788]: INFO nova.scheduler.client.report [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleted allocations for instance f4a490a2-f2b1-4eac-8c1a-a18758583c70 [ 892.190307] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.190814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.191126] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Deleting the datastore file [datastore2] 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.192205] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b895989-0a0d-411f-a56c-47dfa1febf0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.203560] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for the task: (returnval){ [ 892.203560] env[65788]: value = "task-4662566" [ 892.203560] env[65788]: _type = "Task" [ 892.203560] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.214552] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.215212] env[65788]: DEBUG oslo_concurrency.lockutils [None req-194ab18a-f713-46e1-b9e8-5dfe9d384e54 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.089s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.229765] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662562, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.334189] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662563, 'name': CreateVM_Task, 'duration_secs': 0.439561} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.334315] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.334967] env[65788]: WARNING neutronclient.v2_0.client [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 892.335366] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.335525] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 892.335910] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 892.336214] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e6a6abd-03c7-4883-8336-0ae7d63d6284 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.342645] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 892.342645] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520441a7-7c14-ec4a-6362-3988fdf0754d" [ 892.342645] env[65788]: _type = "Task" [ 892.342645] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.353367] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520441a7-7c14-ec4a-6362-3988fdf0754d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.400457] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662564, 'name': PowerOffVM_Task, 'duration_secs': 0.482187} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.400733] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 892.472977] env[65788]: INFO nova.compute.manager [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Detaching volume 0d9578e1-90ca-4aaa-8fac-028b7f80c67a [ 892.498801] env[65788]: DEBUG nova.network.neutron [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Successfully updated port: 987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 892.524624] env[65788]: INFO nova.virt.block_device [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Attempting to driver detach volume 0d9578e1-90ca-4aaa-8fac-028b7f80c67a from mountpoint /dev/sdb [ 892.526177] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 892.526438] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 892.527311] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3505d0-fa92-4dcd-9aed-e247220a887d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.560778] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d00b454-77f1-43f7-92da-2ddaca8eb5df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.565664] env[65788]: DEBUG nova.compute.manager [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Received event network-changed-08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 892.565664] env[65788]: DEBUG nova.compute.manager [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Refreshing instance network info cache due to event network-changed-08be336c-c7bd-448b-b11d-931b7e9788c9. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 892.565664] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Acquiring lock "refresh_cache-7c88e3ef-82d2-46ac-a350-999de8091c01" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.565664] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Acquired lock "refresh_cache-7c88e3ef-82d2-46ac-a350-999de8091c01" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 892.565664] env[65788]: DEBUG nova.network.neutron [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Refreshing network info cache for port 08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 892.566917] env[65788]: INFO nova.compute.manager [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Took 36.44 seconds to build instance. [ 892.574092] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c2ffe0-6513-452c-8314-14b5c0088f37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.607474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99b893b-643e-407e-a0d0-4c0b645f1e2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.628309] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The volume has not been displaced from its original location: [datastore1] volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a/volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 892.633904] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfiguring VM instance instance-00000038 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 892.635412] env[65788]: DEBUG nova.compute.manager [req-e1fd59d6-c876-4734-9d73-0b0518b895b5 req-f487a7c2-0a36-4881-b566-efd3137820df service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Received event network-vif-plugged-987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 892.635616] env[65788]: DEBUG oslo_concurrency.lockutils [req-e1fd59d6-c876-4734-9d73-0b0518b895b5 req-f487a7c2-0a36-4881-b566-efd3137820df service nova] Acquiring lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 892.635820] env[65788]: DEBUG oslo_concurrency.lockutils [req-e1fd59d6-c876-4734-9d73-0b0518b895b5 req-f487a7c2-0a36-4881-b566-efd3137820df service nova] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 892.635979] env[65788]: DEBUG oslo_concurrency.lockutils [req-e1fd59d6-c876-4734-9d73-0b0518b895b5 req-f487a7c2-0a36-4881-b566-efd3137820df service nova] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.636155] env[65788]: DEBUG nova.compute.manager [req-e1fd59d6-c876-4734-9d73-0b0518b895b5 req-f487a7c2-0a36-4881-b566-efd3137820df service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] No waiting events found dispatching network-vif-plugged-987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 892.636319] env[65788]: WARNING nova.compute.manager [req-e1fd59d6-c876-4734-9d73-0b0518b895b5 req-f487a7c2-0a36-4881-b566-efd3137820df service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Received unexpected event network-vif-plugged-987d7302-963a-4e9f-abaf-6d560099bdbc for instance with vm_state building and task_state spawning. [ 892.637175] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4974445c-0fba-400b-b079-534258c3a0c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.657177] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39c9f293-3eb1-48c5-91fd-7d9f011e1d5a tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "f4a490a2-f2b1-4eac-8c1a-a18758583c70" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.041s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.665954] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 892.665954] env[65788]: value = "task-4662567" [ 892.665954] env[65788]: _type = "Task" [ 892.665954] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.682689] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662567, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.712705] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.727038] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662562, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655893} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.727294] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 5e27dbf0-44d2-46a4-87b2-c209afb9559e/5e27dbf0-44d2-46a4-87b2-c209afb9559e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.727508] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.728181] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34d7f094-0ca2-482f-b140-0f8c1948b565 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.738157] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 892.738157] env[65788]: value = "task-4662568" [ 892.738157] env[65788]: _type = "Task" [ 892.738157] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.748391] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662568, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.856939] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.857239] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Processing image e6f1a08c-b319-4a8d-9b85-bd96892a6aab {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.857515] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.857908] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 892.857908] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.858342] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2c5eba53-a933-406f-acaa-006477d92aa5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.868649] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.868843] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.869650] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cadb75c-c2f6-43d3-92ef-d1b739a0aea8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.878856] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 892.878856] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52acda5f-e6e6-6542-d66d-c6940572f6c8" [ 892.878856] env[65788]: _type = "Task" [ 892.878856] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.887700] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52acda5f-e6e6-6542-d66d-c6940572f6c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.002843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.002843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 893.003194] env[65788]: DEBUG nova.network.neutron [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 893.070317] env[65788]: WARNING neutronclient.v2_0.client [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.070422] env[65788]: WARNING openstack [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.071113] env[65788]: WARNING openstack [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.083100] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4c7e159a-ae47-4b23-9433-6f3b543b5822 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.966s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.173474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4f1ca7-6e5e-40ae-862c-417c97f69fad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.180102] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662567, 'name': ReconfigVM_Task, 'duration_secs': 0.270639} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.181104] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfigured VM instance instance-00000038 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 893.188918] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06337b64-2f73-475c-89d4-62b459180fc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.200822] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4c7c09-e726-4fb6-b649-a7a74aa22e4d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.212711] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 893.212711] env[65788]: value = "task-4662569" [ 893.212711] env[65788]: _type = "Task" [ 893.212711] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.247847] env[65788]: DEBUG oslo_vmware.api [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Task: {'id': task-4662566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.578678} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.259115] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e9bf83-454b-4743-808d-5cbba4efb47e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.261929] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.262159] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.262981] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.262981] env[65788]: INFO nova.compute.manager [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Took 1.71 seconds to destroy the instance on the hypervisor. [ 893.262981] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 893.263331] env[65788]: DEBUG nova.compute.manager [-] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 893.263447] env[65788]: DEBUG nova.network.neutron [-] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 893.263693] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.264231] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.264682] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.283688] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cb296e-a3b3-4e61-a966-58d7c9186db5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.288204] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662568, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186311} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.288800] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662569, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.289097] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.290326] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b1acf7-36db-497c-8f7e-cc4f9575cef0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.302537] env[65788]: DEBUG nova.compute.provider_tree [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.306960] env[65788]: WARNING openstack [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.307461] env[65788]: WARNING openstack [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.337435] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 5e27dbf0-44d2-46a4-87b2-c209afb9559e/5e27dbf0-44d2-46a4-87b2-c209afb9559e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.338052] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a26e5fc-efdd-44e0-96fb-522e0c4c3d86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.361880] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 893.361880] env[65788]: value = "task-4662570" [ 893.361880] env[65788]: _type = "Task" [ 893.361880] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.367058] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.376834] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662570, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.391758] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 893.392203] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Fetch image to [datastore1] OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8/OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 893.392500] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Downloading stream optimized image e6f1a08c-b319-4a8d-9b85-bd96892a6aab to [datastore1] OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8/OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8.vmdk on the data store datastore1 as vApp {{(pid=65788) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 893.392703] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Downloading image file data e6f1a08c-b319-4a8d-9b85-bd96892a6aab to the ESX as VM named 'OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8' {{(pid=65788) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 893.489109] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 893.489109] env[65788]: value = "resgroup-9" [ 893.489109] env[65788]: _type = "ResourcePool" [ 893.489109] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 893.489518] env[65788]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d9f8b516-dc2a-472c-bb1a-2500a1228ee0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.507600] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.508129] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.523262] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease: (returnval){ [ 893.523262] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d37-fd9a-52d7-eac4-fa82dafafcfc" [ 893.523262] env[65788]: _type = "HttpNfcLease" [ 893.523262] env[65788]: } obtained for vApp import into resource pool (val){ [ 893.523262] env[65788]: value = "resgroup-9" [ 893.523262] env[65788]: _type = "ResourcePool" [ 893.523262] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 893.523585] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the lease: (returnval){ [ 893.523585] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d37-fd9a-52d7-eac4-fa82dafafcfc" [ 893.523585] env[65788]: _type = "HttpNfcLease" [ 893.523585] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 893.535601] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 893.535601] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d37-fd9a-52d7-eac4-fa82dafafcfc" [ 893.535601] env[65788]: _type = "HttpNfcLease" [ 893.535601] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 893.547872] env[65788]: WARNING neutronclient.v2_0.client [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.550643] env[65788]: WARNING openstack [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.550643] env[65788]: WARNING openstack [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.608042] env[65788]: DEBUG nova.network.neutron [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 893.749131] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.749562] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 893.776264] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662569, 'name': ReconfigVM_Task, 'duration_secs': 0.333347} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.776264] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 893.807376] env[65788]: DEBUG nova.scheduler.client.report [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.814893] env[65788]: DEBUG nova.network.neutron [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Updated VIF entry in instance network info cache for port 08be336c-c7bd-448b-b11d-931b7e9788c9. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 893.815298] env[65788]: DEBUG nova.network.neutron [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Updating instance_info_cache with network_info: [{"id": "08be336c-c7bd-448b-b11d-931b7e9788c9", "address": "fa:16:3e:0b:50:56", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08be336c-c7", "ovs_interfaceid": "08be336c-c7bd-448b-b11d-931b7e9788c9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 893.874805] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662570, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.877758] env[65788]: DEBUG nova.objects.instance [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lazy-loading 'flavor' on Instance uuid d2d614f8-15d9-47d6-9e66-8e1f8fcc18da {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.900520] env[65788]: WARNING neutronclient.v2_0.client [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 893.900520] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 893.900520] env[65788]: WARNING openstack [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.033616] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 894.033616] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d37-fd9a-52d7-eac4-fa82dafafcfc" [ 894.033616] env[65788]: _type = "HttpNfcLease" [ 894.033616] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 894.045352] env[65788]: DEBUG nova.network.neutron [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Updating instance_info_cache with network_info: [{"id": "987d7302-963a-4e9f-abaf-6d560099bdbc", "address": "fa:16:3e:63:ce:63", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap987d7302-96", "ovs_interfaceid": "987d7302-963a-4e9f-abaf-6d560099bdbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 894.216686] env[65788]: DEBUG nova.network.neutron [-] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 894.318740] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.220s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.321879] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d9ba390-3b16-4ae1-b150-efcfe291fcad req-f0294c9e-1392-4db8-8fcf-9367b29344de service nova] Releasing lock "refresh_cache-7c88e3ef-82d2-46ac-a350-999de8091c01" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.322762] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.709s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.323147] env[65788]: DEBUG nova.objects.instance [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lazy-loading 'resources' on Instance uuid bbf6a242-9873-44b9-8938-2f8df11a1018 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.348590] env[65788]: INFO nova.scheduler.client.report [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Deleted allocations for instance 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9 [ 894.372909] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662570, 'name': ReconfigVM_Task, 'duration_secs': 0.523416} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.373219] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 5e27dbf0-44d2-46a4-87b2-c209afb9559e/5e27dbf0-44d2-46a4-87b2-c209afb9559e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.373862] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4fef5547-7cea-470d-8fe9-e192af19a537 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.383202] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 894.383202] env[65788]: value = "task-4662572" [ 894.383202] env[65788]: _type = "Task" [ 894.383202] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.383679] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.383832] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquired lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.383999] env[65788]: DEBUG nova.network.neutron [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 894.384634] env[65788]: DEBUG nova.objects.instance [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lazy-loading 'info_cache' on Instance uuid d2d614f8-15d9-47d6-9e66-8e1f8fcc18da {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.397329] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662572, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.533790] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 894.533790] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d37-fd9a-52d7-eac4-fa82dafafcfc" [ 894.533790] env[65788]: _type = "HttpNfcLease" [ 894.533790] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 894.534032] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 894.534032] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d37-fd9a-52d7-eac4-fa82dafafcfc" [ 894.534032] env[65788]: _type = "HttpNfcLease" [ 894.534032] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 894.534835] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33729312-b7e0-4ff5-970f-c45646b87895 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.543743] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ac394-ca61-a4dd-61b8-b204da529b11/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 894.543936] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ac394-ca61-a4dd-61b8-b204da529b11/disk-0.vmdk. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 894.603740] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.604144] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Instance network_info: |[{"id": "987d7302-963a-4e9f-abaf-6d560099bdbc", "address": "fa:16:3e:63:ce:63", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap987d7302-96", "ovs_interfaceid": "987d7302-963a-4e9f-abaf-6d560099bdbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 894.604642] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:ce:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '987d7302-963a-4e9f-abaf-6d560099bdbc', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.613078] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 894.614073] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.614362] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e904470b-ec00-48c8-b2cc-710b6b487273 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.632602] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "ed55713e-8bf1-4960-8309-47498b163d02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.632602] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.632602] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "ed55713e-8bf1-4960-8309-47498b163d02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.632905] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 894.632905] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.638127] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d5a61e6c-4182-421a-b89c-3cc2fd480b1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.639357] env[65788]: INFO nova.compute.manager [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Terminating instance [ 894.645986] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.645986] env[65788]: value = "task-4662573" [ 894.645986] env[65788]: _type = "Task" [ 894.645986] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.654422] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662573, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.661283] env[65788]: DEBUG nova.compute.manager [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Received event network-changed-987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 894.661283] env[65788]: DEBUG nova.compute.manager [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Refreshing instance network info cache due to event network-changed-987d7302-963a-4e9f-abaf-6d560099bdbc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 894.661451] env[65788]: DEBUG oslo_concurrency.lockutils [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Acquiring lock "refresh_cache-9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.661579] env[65788]: DEBUG oslo_concurrency.lockutils [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Acquired lock "refresh_cache-9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.661720] env[65788]: DEBUG nova.network.neutron [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Refreshing network info cache for port 987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 894.722354] env[65788]: INFO nova.compute.manager [-] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Took 1.46 seconds to deallocate network for instance. [ 894.839476] env[65788]: DEBUG nova.compute.manager [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Received event network-changed-4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 894.839476] env[65788]: DEBUG nova.compute.manager [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Refreshing instance network info cache due to event network-changed-4ac3e7fa-b26e-4239-be70-ae1397fbba52. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 894.839476] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Acquiring lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.840132] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Acquired lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.840477] env[65788]: DEBUG nova.network.neutron [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Refreshing network info cache for port 4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 894.844632] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.845204] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4fcd43a-2821-4838-acde-6e1732d77bca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.860650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1efe6a82-57f3-467b-bfca-c92812c1d2e6 tempest-AttachInterfacesUnderV243Test-2037762512 tempest-AttachInterfacesUnderV243Test-2037762512-project-member] Lock "2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.653s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.864352] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 894.864352] env[65788]: value = "task-4662574" [ 894.864352] env[65788]: _type = "Task" [ 894.864352] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.885523] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 894.885920] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 894.886266] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 894.891250] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518541de-9d53-42ea-960c-17d522a88cd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.903062] env[65788]: DEBUG nova.objects.base [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 894.933734] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662572, 'name': Rename_Task, 'duration_secs': 0.144004} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.941909] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.943710] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9220eaf9-c9cd-44a9-92cc-c04dbc336663 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.952026] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84e931f0-7a36-439e-872e-a4d0e5e1ac81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.963920] env[65788]: WARNING nova.virt.vmwareapi.driver [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 894.964395] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 894.964881] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 894.964881] env[65788]: value = "task-4662575" [ 894.964881] env[65788]: _type = "Task" [ 894.964881] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.966089] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a663261f-33f7-4e36-bdfb-c387d4ec57db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.986834] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.990390] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f50e429-fc6f-42a4-a6a2-5db29d693c67 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.091122] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.091509] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.091602] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.092017] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1678ebef-afb6-4b9e-9eea-aebf42f56b2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.104548] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 895.104548] env[65788]: value = "task-4662577" [ 895.104548] env[65788]: _type = "Task" [ 895.104548] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.116476] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.146101] env[65788]: DEBUG nova.compute.manager [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 895.146216] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.149910] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f80b88-62d3-4099-aa07-8c195629ebbe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.164798] env[65788]: WARNING neutronclient.v2_0.client [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.165592] env[65788]: WARNING openstack [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.166044] env[65788]: WARNING openstack [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.183220] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.183550] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662573, 'name': CreateVM_Task, 'duration_secs': 0.439454} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.188491] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e38b1fcf-6a7d-40f7-8f77-ad872c0c5ce0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.190202] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.191027] env[65788]: WARNING neutronclient.v2_0.client [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.191426] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.191612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 895.192018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 895.192749] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb28e982-a108-479b-9997-bc3bb52b91cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.199800] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 895.199800] env[65788]: value = "task-4662578" [ 895.199800] env[65788]: _type = "Task" [ 895.199800] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.201151] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 895.201151] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52587a5d-ae4c-6b90-b165-e53841631d02" [ 895.201151] env[65788]: _type = "Task" [ 895.201151] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.224907] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662578, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.225267] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52587a5d-ae4c-6b90-b165-e53841631d02, 'name': SearchDatastore_Task, 'duration_secs': 0.011517} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.234661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 895.234936] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.235373] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.235504] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 895.235704] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.236779] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.237477] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-822b4348-f0ed-4761-8d63-a418bb80bcaf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.250976] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.251132] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.257665] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3124e52-fb95-4e31-8f45-b333b1f844f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.268843] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 895.268843] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c53a40-47a6-1841-fbd1-391486f1c0d8" [ 895.268843] env[65788]: _type = "Task" [ 895.268843] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.286313] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c53a40-47a6-1841-fbd1-391486f1c0d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.320745] env[65788]: WARNING openstack [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.320877] env[65788]: WARNING openstack [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.348225] env[65788]: WARNING neutronclient.v2_0.client [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.348225] env[65788]: WARNING openstack [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.348225] env[65788]: WARNING openstack [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.406555] env[65788]: WARNING neutronclient.v2_0.client [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.407457] env[65788]: WARNING openstack [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.407897] env[65788]: WARNING openstack [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.418466] env[65788]: WARNING neutronclient.v2_0.client [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.420641] env[65788]: WARNING openstack [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.420641] env[65788]: WARNING openstack [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.480421] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1dfd124-3143-406c-a7b2-0f1c950c0ed9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.493548] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662575, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.499353] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaed6cc8-01aa-43ce-90de-5b142fb910d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.551016] env[65788]: WARNING openstack [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.551016] env[65788]: WARNING openstack [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.561119] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c894e8c3-3f06-48d7-8aae-d8558275eec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.570990] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba72b5e7-290a-47ee-b2a3-d530528f05f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.586558] env[65788]: DEBUG nova.compute.provider_tree [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.606419] env[65788]: DEBUG nova.network.neutron [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Updated VIF entry in instance network info cache for port 987d7302-963a-4e9f-abaf-6d560099bdbc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 895.606765] env[65788]: DEBUG nova.network.neutron [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Updating instance_info_cache with network_info: [{"id": "987d7302-963a-4e9f-abaf-6d560099bdbc", "address": "fa:16:3e:63:ce:63", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap987d7302-96", "ovs_interfaceid": "987d7302-963a-4e9f-abaf-6d560099bdbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 895.620706] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170278} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.623483] env[65788]: WARNING openstack [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.623845] env[65788]: WARNING openstack [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.630108] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.630330] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 895.630512] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 895.648543] env[65788]: WARNING neutronclient.v2_0.client [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.649265] env[65788]: WARNING openstack [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.649731] env[65788]: WARNING openstack [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.712775] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662578, 'name': PowerOffVM_Task, 'duration_secs': 0.275394} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.713075] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.713268] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.716354] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9ee87ff-235c-480d-8791-0e82164edd11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.729714] env[65788]: WARNING neutronclient.v2_0.client [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 895.730396] env[65788]: WARNING openstack [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.730740] env[65788]: WARNING openstack [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.758851] env[65788]: DEBUG nova.network.neutron [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updated VIF entry in instance network info cache for port 4ac3e7fa-b26e-4239-be70-ae1397fbba52. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 895.759242] env[65788]: DEBUG nova.network.neutron [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [{"id": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "address": "fa:16:3e:88:8c:66", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ac3e7fa-b2", "ovs_interfaceid": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 895.781589] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c53a40-47a6-1841-fbd1-391486f1c0d8, 'name': SearchDatastore_Task, 'duration_secs': 0.015992} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.782495] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84271458-8d58-40a9-8074-a36374637626 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.789518] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 895.789518] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f9da7c-8e02-511c-8412-373bfce60189" [ 895.789518] env[65788]: _type = "Task" [ 895.789518] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.799575] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.799785] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.799961] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleting the datastore file [datastore2] ed55713e-8bf1-4960-8309-47498b163d02 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.801102] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cad16c5e-ce72-48bd-ad7e-f724986d5f8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.808915] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f9da7c-8e02-511c-8412-373bfce60189, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.810442] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for the task: (returnval){ [ 895.810442] env[65788]: value = "task-4662580" [ 895.810442] env[65788]: _type = "Task" [ 895.810442] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.820065] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.864705] env[65788]: DEBUG nova.network.neutron [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Updating instance_info_cache with network_info: [{"id": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "address": "fa:16:3e:e3:e2:2c", "network": {"id": "4ea3a4fc-a85f-41ae-b774-f00355c1fa5e", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-109297227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a84103bf9c94a6bbbb500854a3b9f9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a1140bb-e7", "ovs_interfaceid": "3a1140bb-e791-40d3-a08e-d01e1bfb79b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 895.981481] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662575, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.090394] env[65788]: DEBUG nova.scheduler.client.report [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 896.109056] env[65788]: DEBUG oslo_concurrency.lockutils [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] Releasing lock "refresh_cache-9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.109352] env[65788]: DEBUG nova.compute.manager [req-8963a0f5-5c05-403c-98bd-f6ae18d4a76d req-27ec29c0-1cd1-43bf-9904-0ebbc656422e service nova] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Received event network-vif-deleted-421c2150-8fa1-4607-a700-cba48466f00a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 896.136130] env[65788]: INFO nova.virt.block_device [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Booting with volume 0d9578e1-90ca-4aaa-8fac-028b7f80c67a at /dev/sdb [ 896.173192] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56d789e9-8da1-4bbf-a3bc-bac2c04ffd2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.184160] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bcf237-29c8-4c07-ac94-64b679c9b1b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.227082] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58c2df56-93d4-4f6e-87ab-387d7d638528 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.237589] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027f8ceb-07d6-4ee1-bd34-8764e080768b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.275730] env[65788]: DEBUG oslo_concurrency.lockutils [req-3e010de4-6e75-4069-bb37-34e6cfa03462 req-ba214828-5bed-4fc7-8213-5f391a74a2f3 service nova] Releasing lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.277018] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5b42e8-65e9-44d2-93aa-4438a49e1861 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.285498] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdc29e3-54f4-454f-aa3d-2aabe49ff3cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.303520] env[65788]: DEBUG nova.virt.block_device [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updating existing volume attachment record: 3bda055e-9354-467d-bdb0-b06dafb5bd71 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 896.313510] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f9da7c-8e02-511c-8412-373bfce60189, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.323327] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.368908] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Releasing lock "refresh_cache-d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.480455] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662575, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.596111] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.273s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.600641] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.968s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.600641] env[65788]: DEBUG nova.objects.instance [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lazy-loading 'resources' on Instance uuid 47ff3955-5d9b-4d90-b8b1-276a0ce58a21 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.633478] env[65788]: INFO nova.scheduler.client.report [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Deleted allocations for instance bbf6a242-9873-44b9-8938-2f8df11a1018 [ 896.808987] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f9da7c-8e02-511c-8412-373bfce60189, 'name': SearchDatastore_Task, 'duration_secs': 0.757095} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.808987] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 896.808987] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca/9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.808987] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5f0de1a-a044-4f8c-8940-691f2bb3ac1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.819009] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 896.819009] env[65788]: value = "task-4662581" [ 896.819009] env[65788]: _type = "Task" [ 896.819009] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.826199] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.831555] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662581, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.917987] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 896.918257] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ac394-ca61-a4dd-61b8-b204da529b11/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 896.919653] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c9d22e-ffa8-4f97-9473-1d9b68ed301f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.927928] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ac394-ca61-a4dd-61b8-b204da529b11/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 896.928237] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ac394-ca61-a4dd-61b8-b204da529b11/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 896.928700] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e9fb89c9-c89d-4870-8d8d-01664ee23504 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.984955] env[65788]: DEBUG oslo_vmware.api [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662575, 'name': PowerOnVM_Task, 'duration_secs': 2.007225} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.985456] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.985764] env[65788]: INFO nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Took 10.25 seconds to spawn the instance on the hypervisor. [ 896.986055] env[65788]: DEBUG nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 896.987402] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ca9571-ef04-4012-9bc7-569eb10e7464 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.144767] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6bf9dc81-8f6c-48bd-ae08-c01a2545d905 tempest-ServersWithSpecificFlavorTestJSON-767212478 tempest-ServersWithSpecificFlavorTestJSON-767212478-project-member] Lock "bbf6a242-9873-44b9-8938-2f8df11a1018" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.755s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.209374] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526ac394-ca61-a4dd-61b8-b204da529b11/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 897.209374] env[65788]: INFO nova.virt.vmwareapi.images [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Downloaded image file data e6f1a08c-b319-4a8d-9b85-bd96892a6aab [ 897.211018] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cfd48f-875c-45d0-b41d-c758e3dd698b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.242023] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5068d190-5f0f-4b3e-bb74-ec71e027645c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.331081] env[65788]: DEBUG oslo_vmware.api [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Task: {'id': task-4662580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.146965} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.332322] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.332619] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.332745] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.332914] env[65788]: INFO nova.compute.manager [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Took 2.19 seconds to destroy the instance on the hypervisor. [ 897.333475] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 897.333713] env[65788]: DEBUG nova.compute.manager [-] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 897.333816] env[65788]: DEBUG nova.network.neutron [-] [instance: ed55713e-8bf1-4960-8309-47498b163d02] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 897.334097] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.334737] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 897.334955] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 897.349098] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662581, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.375948] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.375948] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f440b40-ccb7-4077-8296-44d77eaace38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.387029] env[65788]: DEBUG oslo_vmware.api [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 897.387029] env[65788]: value = "task-4662583" [ 897.387029] env[65788]: _type = "Task" [ 897.387029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.387509] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 897.403087] env[65788]: DEBUG oslo_vmware.api [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.509227] env[65788]: INFO nova.virt.vmwareapi.images [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] The imported VM was unregistered [ 897.511611] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 897.511841] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.512434] env[65788]: INFO nova.compute.manager [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Took 40.71 seconds to build instance. [ 897.513386] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e7f9cce-1731-4021-b41f-f597eaf1a34b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.528782] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created directory with path [datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.528987] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8/OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8.vmdk to [datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk. {{(pid=65788) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 897.529882] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-98d94c4d-afe2-41c1-883c-125d3611b370 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.538543] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 897.538543] env[65788]: value = "task-4662584" [ 897.538543] env[65788]: _type = "Task" [ 897.538543] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.552474] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.659565] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579d1c52-321b-4865-84cc-08411e10f9ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.672620] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35076605-eb00-4b42-8521-95b3bf56abd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.707073] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827a8d17-c4f2-432f-9b97-32b4dba94ea0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.717451] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1c47bc-0f49-4127-889a-c7c267652700 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.724581] env[65788]: DEBUG oslo_concurrency.lockutils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.724808] env[65788]: DEBUG oslo_concurrency.lockutils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 897.735043] env[65788]: DEBUG nova.compute.provider_tree [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.737974] env[65788]: DEBUG nova.compute.utils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 897.834448] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662581, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.676855} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.835036] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca/9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.835036] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.835230] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3540f30c-63cf-44dd-838e-8c79fb277d1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.844911] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 897.844911] env[65788]: value = "task-4662585" [ 897.844911] env[65788]: _type = "Task" [ 897.844911] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.858095] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662585, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.900941] env[65788]: DEBUG oslo_vmware.api [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662583, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.926732] env[65788]: DEBUG nova.compute.manager [req-d4c25101-9999-44bd-a30c-a8f44faa56cc req-29dd0bd3-f704-4730-8eda-acc406f9bc37 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Received event network-vif-deleted-ffa53d50-ad0e-438c-994a-b0b7a97f85f2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 897.926936] env[65788]: INFO nova.compute.manager [req-d4c25101-9999-44bd-a30c-a8f44faa56cc req-29dd0bd3-f704-4730-8eda-acc406f9bc37 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Neutron deleted interface ffa53d50-ad0e-438c-994a-b0b7a97f85f2; detaching it from the instance and deleting it from the info cache [ 897.927553] env[65788]: DEBUG nova.network.neutron [req-d4c25101-9999-44bd-a30c-a8f44faa56cc req-29dd0bd3-f704-4730-8eda-acc406f9bc37 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.017127] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6dfdfb8f-98b0-4647-97af-5f46c1519f89 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.224s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.052746] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.146138] env[65788]: DEBUG nova.network.neutron [-] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.241311] env[65788]: DEBUG nova.scheduler.client.report [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.246538] env[65788]: DEBUG oslo_concurrency.lockutils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 0.522s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.357067] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662585, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080396} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.357390] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.358303] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46af88f-f08d-4a4c-af6a-86e0811dc46a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.384188] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca/9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.384612] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a5d1ee9-7599-46d7-baca-eda62c0bd077 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.413891] env[65788]: DEBUG oslo_vmware.api [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662583, 'name': PowerOnVM_Task, 'duration_secs': 0.55555} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.414186] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.414486] env[65788]: DEBUG nova.compute.manager [None req-d752755e-9f91-4df4-bc86-180e6b5d6f05 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 898.415384] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765f75d9-0d55-4e07-a0d4-a2d3ded37374 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.420454] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 898.420454] env[65788]: value = "task-4662586" [ 898.420454] env[65788]: _type = "Task" [ 898.420454] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.438886] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd9172c8-338a-4460-81f8-16793bd5361f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.442048] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662586, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.452595] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dc1a19-d42b-4576-890f-916f5f7f0e17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.482497] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 898.483651] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 898.483651] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 898.483651] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 898.483836] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 898.483933] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 898.484375] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.484613] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 898.484837] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 898.484986] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 898.485206] env[65788]: DEBUG nova.virt.hardware [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 898.487058] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d234e88b-a865-45e9-b119-3ca31ec014a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.505241] env[65788]: DEBUG nova.compute.manager [req-d4c25101-9999-44bd-a30c-a8f44faa56cc req-29dd0bd3-f704-4730-8eda-acc406f9bc37 service nova] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Detach interface failed, port_id=ffa53d50-ad0e-438c-994a-b0b7a97f85f2, reason: Instance ed55713e-8bf1-4960-8309-47498b163d02 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 898.517484] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9b1c27-707e-46f9-8cb7-01a978acd01f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.536932] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:8c:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31daf1d5-0a84-4d17-a694-a7ab16e9b81b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.545395] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 898.546286] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 898.550281] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-569ad6f3-cd79-4f6f-80af-cda736d4fea5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.574681] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.576352] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.576352] env[65788]: value = "task-4662587" [ 898.576352] env[65788]: _type = "Task" [ 898.576352] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.587442] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662587, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.649871] env[65788]: INFO nova.compute.manager [-] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Took 1.32 seconds to deallocate network for instance. [ 898.748053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.750416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.958s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.752027] env[65788]: INFO nova.compute.claims [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.781487] env[65788]: INFO nova.scheduler.client.report [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Deleted allocations for instance 47ff3955-5d9b-4d90-b8b1-276a0ce58a21 [ 898.935390] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662586, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.062019] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.092690] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662587, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.159060] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.290714] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6ecd9046-2d96-41e7-afaf-a7dd0fe25272 tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "47ff3955-5d9b-4d90-b8b1-276a0ce58a21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.081s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.364381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.364734] env[65788]: DEBUG oslo_concurrency.lockutils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.365082] env[65788]: INFO nova.compute.manager [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Attaching volume a5111781-1b06-4e45-b9ac-40d76784a6cc to /dev/sdb [ 899.407448] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccea660b-79d4-4085-8db3-6fc01454356c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.414322] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.414716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.418026] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.418026] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.418026] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.419930] env[65788]: INFO nova.compute.manager [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Terminating instance [ 899.423035] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7601e8c2-cac4-4c30-a9dc-2c2d0d7dda41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.440392] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662586, 'name': ReconfigVM_Task, 'duration_secs': 0.662814} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.444610] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca/9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.445761] env[65788]: DEBUG nova.virt.block_device [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating existing volume attachment record: d8f85ea6-75cf-4d79-ba84-0b1db46311d8 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 899.448312] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa498b65-131a-4f40-8d2d-ba0e865cfcd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.456846] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 899.456846] env[65788]: value = "task-4662588" [ 899.456846] env[65788]: _type = "Task" [ 899.456846] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.468768] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662588, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.559703] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.591175] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662587, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.754049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "2c24bc30-d413-4714-81c2-b657cafe94bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.754049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.757865] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "2c24bc30-d413-4714-81c2-b657cafe94bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.757865] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.757865] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.759852] env[65788]: INFO nova.compute.manager [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Terminating instance [ 899.935094] env[65788]: DEBUG nova.compute.manager [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 899.935349] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.936676] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5feed8a-b758-48e3-b20e-4f0abeadd01f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.946842] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.951171] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41c1c957-3a3b-4251-b8b6-d9730d960a37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.967632] env[65788]: DEBUG oslo_vmware.api [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 899.967632] env[65788]: value = "task-4662592" [ 899.967632] env[65788]: _type = "Task" [ 899.967632] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.980413] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662588, 'name': Rename_Task, 'duration_secs': 0.26446} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.980413] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.980494] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f83559db-543a-4995-a7cb-539ec38f60d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.987405] env[65788]: DEBUG oslo_vmware.api [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662592, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.998538] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 899.998538] env[65788]: value = "task-4662593" [ 899.998538] env[65788]: _type = "Task" [ 899.998538] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.013391] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662593, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.065760] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.098072] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662587, 'name': CreateVM_Task, 'duration_secs': 1.231317} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.098859] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 900.099240] env[65788]: WARNING neutronclient.v2_0.client [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 900.099721] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.101987] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 900.101987] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 900.101987] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35635169-a395-45f9-8372-6d7209773a24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.110797] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 900.110797] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521b46d1-ce6f-2925-2a43-6270e1e2386a" [ 900.110797] env[65788]: _type = "Task" [ 900.110797] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.126143] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521b46d1-ce6f-2925-2a43-6270e1e2386a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.269028] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8894f15e-b3bc-47b5-bbf2-3c8ba98f05e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.272284] env[65788]: DEBUG nova.compute.manager [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 900.272527] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.273419] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9534cc8e-2ecf-4c70-a659-b1527f04159a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.285432] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45623707-177b-4766-bde1-46f4b18ed954 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.289250] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.289530] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d528833d-6d40-4f28-b711-409d4703b9d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.324813] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eadb7af-ca51-474b-a751-e7163c92444b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.328533] env[65788]: DEBUG oslo_vmware.api [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 900.328533] env[65788]: value = "task-4662594" [ 900.328533] env[65788]: _type = "Task" [ 900.328533] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.338405] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c564558b-856c-4c98-beda-9d353c13db8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.348493] env[65788]: DEBUG oslo_vmware.api [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.361567] env[65788]: DEBUG nova.compute.provider_tree [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.478206] env[65788]: DEBUG oslo_vmware.api [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662592, 'name': PowerOffVM_Task, 'duration_secs': 0.503408} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.478550] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.478739] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.479108] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c22f0c6-cdf1-4802-977f-a486dbe1d05e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.509869] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662593, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.555354] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.555597] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.555779] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleting the datastore file [datastore1] 5e27dbf0-44d2-46a4-87b2-c209afb9559e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.556548] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9218b4dd-76ca-4323-a0d1-061a3d637b44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.562015] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662584, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.855049} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.562767] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8/OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8.vmdk to [datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk. [ 900.562928] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Cleaning up location [datastore1] OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8 {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 900.563116] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_c7bb3019-5e77-4264-8dbb-f0da22b01fd8 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.563400] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e93bc3e4-46a4-4b88-9759-602c273f55f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.567106] env[65788]: DEBUG oslo_vmware.api [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 900.567106] env[65788]: value = "task-4662596" [ 900.567106] env[65788]: _type = "Task" [ 900.567106] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.571373] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 900.571373] env[65788]: value = "task-4662597" [ 900.571373] env[65788]: _type = "Task" [ 900.571373] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.580153] env[65788]: DEBUG oslo_vmware.api [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.590332] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662597, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.622197] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521b46d1-ce6f-2925-2a43-6270e1e2386a, 'name': SearchDatastore_Task, 'duration_secs': 0.052123} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.623156] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 900.623156] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.623156] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.623156] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 900.623535] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.623701] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c70272f-e738-44ef-8942-6da0982d9854 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.633451] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.633603] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 900.634688] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5af882a-2c17-4f43-90b9-5e82234e7758 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.640597] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 900.640597] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5203a0c5-974b-3bf6-5105-ad1208f12ef4" [ 900.640597] env[65788]: _type = "Task" [ 900.640597] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.649060] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5203a0c5-974b-3bf6-5105-ad1208f12ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.840467] env[65788]: DEBUG oslo_vmware.api [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662594, 'name': PowerOffVM_Task, 'duration_secs': 0.310819} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.840747] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.840901] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.841190] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1d68199-efbe-48a7-bf0f-3fc53ab9c0bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.866189] env[65788]: DEBUG nova.scheduler.client.report [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 900.928594] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.928922] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.929126] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Deleting the datastore file [datastore2] 2c24bc30-d413-4714-81c2-b657cafe94bd {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.929406] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e449067-81bd-40d5-b759-4392643c6d2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.937441] env[65788]: DEBUG oslo_vmware.api [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for the task: (returnval){ [ 900.937441] env[65788]: value = "task-4662599" [ 900.937441] env[65788]: _type = "Task" [ 900.937441] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.945988] env[65788]: DEBUG oslo_vmware.api [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662599, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.009979] env[65788]: DEBUG oslo_vmware.api [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662593, 'name': PowerOnVM_Task, 'duration_secs': 0.828932} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.010364] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.010595] env[65788]: INFO nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Took 9.03 seconds to spawn the instance on the hypervisor. [ 901.010803] env[65788]: DEBUG nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 901.011659] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee6ac96-e3e0-4c2f-a4d6-6823b2cd4820 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.080114] env[65788]: DEBUG oslo_vmware.api [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163214} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.080860] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.081064] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.081251] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.081449] env[65788]: INFO nova.compute.manager [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 901.081934] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 901.082416] env[65788]: DEBUG nova.compute.manager [-] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 901.082554] env[65788]: DEBUG nova.network.neutron [-] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 901.082842] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.083554] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.083857] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.096417] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662597, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038546} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.096754] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.097045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.097335] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk to [datastore1] 7c88e3ef-82d2-46ac-a350-999de8091c01/7c88e3ef-82d2-46ac-a350-999de8091c01.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.097634] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-697c3131-4099-4777-a153-191370944ae5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.105761] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 901.105761] env[65788]: value = "task-4662600" [ 901.105761] env[65788]: _type = "Task" [ 901.105761] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.114941] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.152810] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5203a0c5-974b-3bf6-5105-ad1208f12ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.010269} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.153699] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f309b0f7-9cbd-4f79-808f-930a4f6f67cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.157137] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.163901] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 901.163901] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528b71e2-586a-0aec-96fd-b11e13e4c113" [ 901.163901] env[65788]: _type = "Task" [ 901.163901] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.174704] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528b71e2-586a-0aec-96fd-b11e13e4c113, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.375025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.375025] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 901.376256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.115s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.376693] env[65788]: DEBUG nova.objects.instance [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lazy-loading 'resources' on Instance uuid d3e2b5f6-20f1-40ac-b92e-e110ee119178 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.454480] env[65788]: DEBUG oslo_vmware.api [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Task: {'id': task-4662599, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150046} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.455304] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.455847] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.457035] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.457035] env[65788]: INFO nova.compute.manager [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Took 1.18 seconds to destroy the instance on the hypervisor. [ 901.457861] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 901.458260] env[65788]: DEBUG nova.compute.manager [-] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 901.459081] env[65788]: DEBUG nova.network.neutron [-] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 901.459891] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.461163] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.461841] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.532326] env[65788]: INFO nova.compute.manager [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Took 39.10 seconds to build instance. [ 901.621390] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.649377] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.681890] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528b71e2-586a-0aec-96fd-b11e13e4c113, 'name': SearchDatastore_Task, 'duration_secs': 0.011566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.682329] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.682623] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.682954] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4239b62b-d5e4-4775-beb4-b4135d7b9610 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.695685] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 901.695685] env[65788]: value = "task-4662601" [ 901.695685] env[65788]: _type = "Task" [ 901.695685] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.708434] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.880623] env[65788]: DEBUG nova.compute.utils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 901.885986] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 901.885986] env[65788]: DEBUG nova.network.neutron [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 901.885986] env[65788]: WARNING neutronclient.v2_0.client [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.886446] env[65788]: WARNING neutronclient.v2_0.client [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 901.886845] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.887214] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.958203] env[65788]: DEBUG nova.network.neutron [-] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 902.006201] env[65788]: DEBUG nova.policy [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdb591e14bcf4170b720e9e82b9d8f15', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '483bc5fc16c443ab89591367b0a588ab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 902.036797] env[65788]: DEBUG oslo_concurrency.lockutils [None req-20d7ca54-ee68-4020-844b-47b0d6915d26 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.617s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 902.078266] env[65788]: DEBUG nova.compute.manager [req-899b50c7-1bd0-4d9d-8eb6-3b05e1178a52 req-98b21acd-37ea-44f5-b8bf-2dbce65ddd89 service nova] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Received event network-vif-deleted-d6457e25-41bd-486e-bd24-5ec2f17cfbb3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 902.128413] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.209626] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.386362] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 902.406460] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5216313b-d696-4b1b-915f-86481f60f58d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.420942] env[65788]: DEBUG nova.network.neutron [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Successfully created port: 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 902.425852] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b898ec32-5743-43a2-9870-8fcaf933e527 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.471324] env[65788]: DEBUG nova.network.neutron [-] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 902.474267] env[65788]: INFO nova.compute.manager [-] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Took 1.39 seconds to deallocate network for instance. [ 902.477053] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37f0664-90a1-4a1a-a304-eb5a89c9a797 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.495103] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ac37a8-1e52-4aba-8571-955bee736438 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.513133] env[65788]: DEBUG nova.compute.provider_tree [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.620801] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.711997] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662601, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.975856] env[65788]: INFO nova.compute.manager [-] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Took 1.52 seconds to deallocate network for instance. [ 902.984809] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.017859] env[65788]: DEBUG nova.scheduler.client.report [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.123107] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.212496] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662601, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.404350] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 903.435430] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 903.435723] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 903.436069] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 903.436236] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 903.436400] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 903.436548] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 903.436759] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.436913] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 903.437198] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 903.437257] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 903.437451] env[65788]: DEBUG nova.virt.hardware [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 903.438738] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43343191-043e-4263-8eef-e18608daef71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.452710] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99767347-86f9-4f6c-b401-6bc74e780cee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.483121] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.524584] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.527614] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.602s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 903.530743] env[65788]: INFO nova.compute.claims [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.556851] env[65788]: INFO nova.scheduler.client.report [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted allocations for instance d3e2b5f6-20f1-40ac-b92e-e110ee119178 [ 903.623429] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.711239] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662601, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.991826} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.711590] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.711842] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.712164] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f16fcdc-3d35-49e8-ae27-903c4326bcd9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.723621] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 903.723621] env[65788]: value = "task-4662603" [ 903.723621] env[65788]: _type = "Task" [ 903.723621] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.735579] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662603, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.011776] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 904.012380] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910320', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'name': 'volume-a5111781-1b06-4e45-b9ac-40d76784a6cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3a06a833-5aaa-4b5d-88b3-8a1d469580af', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'serial': 'a5111781-1b06-4e45-b9ac-40d76784a6cc'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 904.012923] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6972c9fd-89ae-40f0-a83b-c573fc7268bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.031906] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf35d7dc-f0a8-4699-b8be-f0239acac93e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.066166] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] volume-a5111781-1b06-4e45-b9ac-40d76784a6cc/volume-a5111781-1b06-4e45-b9ac-40d76784a6cc.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.067929] env[65788]: DEBUG nova.network.neutron [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Successfully updated port: 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 904.072237] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05e9b1d4-06db-4a86-8ae3-ef11e0b125d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.092726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.092918] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.093114] env[65788]: DEBUG nova.network.neutron [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 904.094837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f0b0f716-4fbc-4ad6-8332-ee5e12c3b08c tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "d3e2b5f6-20f1-40ac-b92e-e110ee119178" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.989s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.104867] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 904.104867] env[65788]: value = "task-4662604" [ 904.104867] env[65788]: _type = "Task" [ 904.104867] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.123108] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.127367] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662604, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.237834] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076275} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.237938] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.239544] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e24b6d-48a8-421c-827d-e80e3c918450 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.267480] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.268233] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5a9bee1-13b5-4c72-b424-c8113a3c3577 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.290485] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 904.290485] env[65788]: value = "task-4662605" [ 904.290485] env[65788]: _type = "Task" [ 904.290485] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.300229] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662605, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.507019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.507019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.597909] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.598328] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.619255] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662604, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.628877] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.638057] env[65788]: DEBUG nova.network.neutron [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 904.667566] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.667985] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.743713] env[65788]: WARNING neutronclient.v2_0.client [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 904.744312] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 904.744631] env[65788]: WARNING openstack [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 904.802865] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.814623] env[65788]: DEBUG nova.compute.manager [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Received event network-vif-deleted-05d3722d-6c89-4066-84f7-83e285df6855 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 904.814802] env[65788]: DEBUG nova.compute.manager [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-vif-plugged-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 904.814992] env[65788]: DEBUG oslo_concurrency.lockutils [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Acquiring lock "b4f157a7-350c-4fff-8509-44426714846c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.815222] env[65788]: DEBUG oslo_concurrency.lockutils [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Lock "b4f157a7-350c-4fff-8509-44426714846c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.815391] env[65788]: DEBUG oslo_concurrency.lockutils [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Lock "b4f157a7-350c-4fff-8509-44426714846c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.815553] env[65788]: DEBUG nova.compute.manager [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] No waiting events found dispatching network-vif-plugged-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 904.815713] env[65788]: WARNING nova.compute.manager [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received unexpected event network-vif-plugged-891ac553-14f8-41d2-a6e4-c24ef3c62c6a for instance with vm_state building and task_state spawning. [ 904.815869] env[65788]: DEBUG nova.compute.manager [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 904.816180] env[65788]: DEBUG nova.compute.manager [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing instance network info cache due to event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 904.816309] env[65788]: DEBUG oslo_concurrency.lockutils [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.864059] env[65788]: DEBUG nova.network.neutron [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 904.987318] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bb1472-42c1-4aff-a7e2-f13b21eba26c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.999199] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14cb5ff-c5a0-4aac-902a-ff8e6aa0aabd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.041540] env[65788]: DEBUG nova.compute.utils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 905.045060] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52883704-df07-434c-861f-34fef40d6449 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.055053] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b707f42-bb1c-425c-ad62-255af82efac8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.071651] env[65788]: DEBUG nova.compute.provider_tree [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.123697] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662604, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.130128] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662600, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.742428} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.130466] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e6f1a08c-b319-4a8d-9b85-bd96892a6aab/e6f1a08c-b319-4a8d-9b85-bd96892a6aab.vmdk to [datastore1] 7c88e3ef-82d2-46ac-a350-999de8091c01/7c88e3ef-82d2-46ac-a350-999de8091c01.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.131389] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbe75e5-69ca-4753-88e5-63d3fdff7cd7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.159498] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 7c88e3ef-82d2-46ac-a350-999de8091c01/7c88e3ef-82d2-46ac-a350-999de8091c01.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.159498] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f3cfc9a-62b6-41d6-ba14-5ab1d972af5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.185570] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 905.185570] env[65788]: value = "task-4662606" [ 905.185570] env[65788]: _type = "Task" [ 905.185570] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.195036] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662606, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.303981] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662605, 'name': ReconfigVM_Task, 'duration_secs': 0.893156} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.304468] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfigured VM instance instance-00000038 to attach disk [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1/f160e4a5-1e91-495e-800e-116ec435d8e1.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.306102] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'size': 0, 'boot_index': 0, 'disk_bus': None, 'device_type': 'disk', 'encrypted': False, 'encryption_secret_uuid': None, 'guest_format': None, 'encryption_options': None, 'encryption_format': None, 'image_id': 'd5a7b45a-a1f3-4d64-bb88-d3ab995bc18b'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '3bda055e-9354-467d-bdb0-b06dafb5bd71', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'}, 'disk_bus': None, 'device_type': None, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65788) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 905.306356] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 905.306597] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 905.307494] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c446ad-5f49-45ad-b924-4bc30fc5867a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.325495] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a208c1c1-6fc3-481e-b361-16d7fcfdfb14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.352316] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a/volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.353076] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5c2eeb2-eb1b-489d-adff-1ac4a13c1902 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.367391] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.367768] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Instance network_info: |[{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 905.368147] env[65788]: DEBUG oslo_concurrency.lockutils [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.368436] env[65788]: DEBUG nova.network.neutron [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 905.370337] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:58:21:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4162774e-ec80-4d85-aeb4-fae77f197393', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '891ac553-14f8-41d2-a6e4-c24ef3c62c6a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.379290] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Creating folder: Project (483bc5fc16c443ab89591367b0a588ab). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.381422] env[65788]: WARNING neutronclient.v2_0.client [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 905.382089] env[65788]: WARNING openstack [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.382580] env[65788]: WARNING openstack [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.389986] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-596d3431-da9d-4a34-9849-b9c1afedf6b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.393732] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 905.393732] env[65788]: value = "task-4662607" [ 905.393732] env[65788]: _type = "Task" [ 905.393732] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.403541] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662607, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.406046] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Created folder: Project (483bc5fc16c443ab89591367b0a588ab) in parent group-v910111. [ 905.406046] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Creating folder: Instances. Parent ref: group-v910321. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.406046] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34830a86-7f94-42a0-9db9-8af8f7183fe4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.419077] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Created folder: Instances in parent group-v910321. [ 905.420868] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 905.420868] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4f157a7-350c-4fff-8509-44426714846c] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.420868] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f0b5f8ab-5f68-4b4f-abd2-bcf57d7f7ac6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.443690] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.443690] env[65788]: value = "task-4662610" [ 905.443690] env[65788]: _type = "Task" [ 905.443690] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.447616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "5cef20ed-d9f0-4237-ae9d-da401b953904" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.447872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.448108] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "5cef20ed-d9f0-4237-ae9d-da401b953904-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.448340] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.448521] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.450815] env[65788]: INFO nova.compute.manager [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Terminating instance [ 905.456876] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662610, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.501040] env[65788]: WARNING openstack [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.501464] env[65788]: WARNING openstack [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.545647] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.039s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.555505] env[65788]: WARNING neutronclient.v2_0.client [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 905.556316] env[65788]: WARNING openstack [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.556726] env[65788]: WARNING openstack [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.575837] env[65788]: DEBUG nova.scheduler.client.report [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.622305] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662604, 'name': ReconfigVM_Task, 'duration_secs': 1.33028} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.623172] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfigured VM instance instance-0000003b to attach disk [datastore2] volume-a5111781-1b06-4e45-b9ac-40d76784a6cc/volume-a5111781-1b06-4e45-b9ac-40d76784a6cc.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.627886] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da87282d-3fa8-404e-9db1-cfb6821aee78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.647585] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 905.647585] env[65788]: value = "task-4662611" [ 905.647585] env[65788]: _type = "Task" [ 905.647585] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.655348] env[65788]: DEBUG nova.network.neutron [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updated VIF entry in instance network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 905.655840] env[65788]: DEBUG nova.network.neutron [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 905.662667] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.697993] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662606, 'name': ReconfigVM_Task, 'duration_secs': 0.305446} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.698327] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 7c88e3ef-82d2-46ac-a350-999de8091c01/7c88e3ef-82d2-46ac-a350-999de8091c01.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.699159] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbd61d11-65e2-4565-ba49-9052e6a38ccc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.707725] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 905.707725] env[65788]: value = "task-4662612" [ 905.707725] env[65788]: _type = "Task" [ 905.707725] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.719604] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662612, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.904250] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662607, 'name': ReconfigVM_Task, 'duration_secs': 0.323789} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.904397] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfigured VM instance instance-00000038 to attach disk [datastore1] volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a/volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.909099] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-294a1ea5-06e6-4316-880f-34139d4805ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.926017] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 905.926017] env[65788]: value = "task-4662613" [ 905.926017] env[65788]: _type = "Task" [ 905.926017] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.935082] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662613, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.954505] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662610, 'name': CreateVM_Task, 'duration_secs': 0.433245} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.954907] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4f157a7-350c-4fff-8509-44426714846c] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.955432] env[65788]: WARNING neutronclient.v2_0.client [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 905.955916] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.956151] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.956585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 905.956926] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3281faf-14a9-40cf-9944-ab4f37686914 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.959932] env[65788]: DEBUG nova.compute.manager [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 905.960247] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 905.961546] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b070a5-7306-4a8b-8df0-2ecfb1040b44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.970719] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.971601] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96489dd6-7d6a-4792-ad04-6563e614909a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.973514] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 905.973514] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f38c5a-3de4-c0be-4e1f-626a4e46a187" [ 905.973514] env[65788]: _type = "Task" [ 905.973514] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.980789] env[65788]: DEBUG oslo_vmware.api [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 905.980789] env[65788]: value = "task-4662614" [ 905.980789] env[65788]: _type = "Task" [ 905.980789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.988560] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f38c5a-3de4-c0be-4e1f-626a4e46a187, 'name': SearchDatastore_Task, 'duration_secs': 0.011551} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.989632] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 905.990041] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.990408] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.990663] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.990915] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.991545] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c52d66ba-db7b-4dce-a99a-ab692d8697d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.997833] env[65788]: DEBUG oslo_vmware.api [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662614, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.005555] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.005778] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.006647] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a185aaf-f62c-4135-ac11-7ffa92cf5099 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.013661] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 906.013661] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52922a58-8617-511f-1f38-499daf72fd0c" [ 906.013661] env[65788]: _type = "Task" [ 906.013661] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.023867] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52922a58-8617-511f-1f38-499daf72fd0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.081444] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.081948] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 906.085824] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.328s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.085824] env[65788]: DEBUG nova.objects.instance [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lazy-loading 'resources' on Instance uuid 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.159925] env[65788]: DEBUG oslo_concurrency.lockutils [req-edb5860b-843f-4b67-a6c0-53fc7b824d7c req-66d3fc4c-aa8d-4f48-b0de-f01b7ea2b6c0 service nova] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.160645] env[65788]: DEBUG oslo_vmware.api [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662611, 'name': ReconfigVM_Task, 'duration_secs': 0.215931} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.161080] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910320', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'name': 'volume-a5111781-1b06-4e45-b9ac-40d76784a6cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3a06a833-5aaa-4b5d-88b3-8a1d469580af', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'serial': 'a5111781-1b06-4e45-b9ac-40d76784a6cc'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 906.226070] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662612, 'name': Rename_Task, 'duration_secs': 0.173596} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.226643] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.226913] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-896cbf39-ff78-49a1-b896-9727dbebdb6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.236200] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 906.236200] env[65788]: value = "task-4662615" [ 906.236200] env[65788]: _type = "Task" [ 906.236200] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.247490] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662615, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.436550] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662613, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.491253] env[65788]: DEBUG oslo_vmware.api [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662614, 'name': PowerOffVM_Task, 'duration_secs': 0.300996} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.491540] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.491699] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.492407] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78f67311-0a4f-469d-aea6-b500ea4eaa54 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.525810] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52922a58-8617-511f-1f38-499daf72fd0c, 'name': SearchDatastore_Task, 'duration_secs': 0.012125} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.527039] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a0ba624-18d3-4dd7-8ba2-8db87d8453d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.532730] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 906.532730] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271c059-3cc5-8431-2dd2-ca5d67f3a5d2" [ 906.532730] env[65788]: _type = "Task" [ 906.532730] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.542020] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271c059-3cc5-8431-2dd2-ca5d67f3a5d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.576992] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.577367] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.577589] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleting the datastore file [datastore2] 5cef20ed-d9f0-4237-ae9d-da401b953904 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.577966] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a80610cc-bad0-42da-a09d-09f34778e4f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.581220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.581537] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.581866] env[65788]: INFO nova.compute.manager [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Attaching volume 249d7aab-deda-4273-9967-6fde28af77b4 to /dev/sdb [ 906.588386] env[65788]: DEBUG nova.compute.utils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 906.598714] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 906.599051] env[65788]: DEBUG nova.network.neutron [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 906.599559] env[65788]: WARNING neutronclient.v2_0.client [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.600013] env[65788]: WARNING neutronclient.v2_0.client [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 906.600992] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.601614] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.613256] env[65788]: DEBUG oslo_vmware.api [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for the task: (returnval){ [ 906.613256] env[65788]: value = "task-4662617" [ 906.613256] env[65788]: _type = "Task" [ 906.613256] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.632142] env[65788]: DEBUG oslo_vmware.api [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662617, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.643892] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837fcaf5-9644-4660-ac44-9714aea5a652 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.655200] env[65788]: DEBUG nova.policy [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '105258eea35d4e22aa4914ec5b5ccbdf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22a3db3230244ef5af1227df6f4a6f12', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 906.659911] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de372760-d25a-4192-9b74-c764f4b1ceca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.683021] env[65788]: DEBUG nova.virt.block_device [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Updating existing volume attachment record: 5b3ed5ea-121b-46a8-8b0b-6e4817ccb234 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 906.732139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.732646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.732880] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.733561] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.733668] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.736682] env[65788]: INFO nova.compute.manager [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Terminating instance [ 906.751688] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662615, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.942341] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662613, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.031742] env[65788]: DEBUG nova.network.neutron [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Successfully created port: 85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 907.052485] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271c059-3cc5-8431-2dd2-ca5d67f3a5d2, 'name': SearchDatastore_Task, 'duration_secs': 0.010425} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.053104] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.053539] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b4f157a7-350c-4fff-8509-44426714846c/b4f157a7-350c-4fff-8509-44426714846c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.053896] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa9108fc-0b4f-436f-b5d3-1bc8c0410d59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.059464] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f9f349c-1437-4aaa-bfc6-122822f64439 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.064772] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 907.064772] env[65788]: value = "task-4662619" [ 907.064772] env[65788]: _type = "Task" [ 907.064772] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.073323] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda9fa74-4a04-4b7a-8220-fe99daefc9da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.081636] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.113194] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 907.118658] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71b5e33-5463-4524-86cd-89910f733531 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.134865] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa160f55-3a38-4e6d-91d0-212ae9a4b79d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.138547] env[65788]: DEBUG oslo_vmware.api [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Task: {'id': task-4662617, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169438} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.138999] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.138999] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.139182] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.140805] env[65788]: INFO nova.compute.manager [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Took 1.18 seconds to destroy the instance on the hypervisor. [ 907.140805] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 907.140805] env[65788]: DEBUG nova.compute.manager [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 907.140805] env[65788]: DEBUG nova.network.neutron [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 907.140805] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 907.141241] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.141500] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.158651] env[65788]: DEBUG nova.compute.provider_tree [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.214687] env[65788]: DEBUG nova.objects.instance [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'flavor' on Instance uuid 3a06a833-5aaa-4b5d-88b3-8a1d469580af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.246146] env[65788]: DEBUG nova.compute.manager [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 907.246456] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 907.247329] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9647f2f1-8082-44ff-a5df-a539e419749c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.259213] env[65788]: DEBUG oslo_vmware.api [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662615, 'name': PowerOnVM_Task, 'duration_secs': 0.773128} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.263073] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.263422] env[65788]: INFO nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Took 17.85 seconds to spawn the instance on the hypervisor. [ 907.263464] env[65788]: DEBUG nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 907.263816] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.265149] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6dd685-184a-4fc9-8803-18ac0a398cb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.268094] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c9d4d74-41e8-4b7d-86c4-fa9309452612 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.282499] env[65788]: DEBUG oslo_vmware.api [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 907.282499] env[65788]: value = "task-4662621" [ 907.282499] env[65788]: _type = "Task" [ 907.282499] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.287334] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 907.443202] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662613, 'name': ReconfigVM_Task, 'duration_secs': 1.19495} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.443944] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 907.444477] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b17dc19c-0af9-412e-9854-b5663c0ad869 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.457609] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 907.457609] env[65788]: value = "task-4662623" [ 907.457609] env[65788]: _type = "Task" [ 907.457609] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.478174] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662623, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.577882] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662619, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.662888] env[65788]: DEBUG nova.scheduler.client.report [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.718929] env[65788]: DEBUG oslo_concurrency.lockutils [None req-549a2cab-41bb-428d-a7bf-d6f1f8e29d15 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.354s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.796695] env[65788]: INFO nova.compute.manager [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Took 50.94 seconds to build instance. [ 907.803704] env[65788]: DEBUG oslo_vmware.api [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662621, 'name': PowerOffVM_Task, 'duration_secs': 0.44495} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.803945] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.804127] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.804391] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2555236-37c8-4e1a-8bf6-6d693188444e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.877275] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.877528] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.877602] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleting the datastore file [datastore2] 8d7a33e6-08fb-481e-8fac-fdf13b19aa21 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.878574] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6e59e0e3-cabe-417b-b03e-71a6f63e5049 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.886736] env[65788]: DEBUG oslo_vmware.api [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 907.886736] env[65788]: value = "task-4662625" [ 907.886736] env[65788]: _type = "Task" [ 907.886736] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.898698] env[65788]: DEBUG oslo_vmware.api [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662625, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.970761] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662623, 'name': Rename_Task, 'duration_secs': 0.297169} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.971095] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.971980] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba5e7bfd-799b-4f44-95ab-eeb904667c6a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.982789] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 907.982789] env[65788]: value = "task-4662626" [ 907.982789] env[65788]: _type = "Task" [ 907.982789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.996487] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662626, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.078320] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662619, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541673} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.078638] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b4f157a7-350c-4fff-8509-44426714846c/b4f157a7-350c-4fff-8509-44426714846c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.078848] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.079167] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69831dac-1a00-4e5f-b83a-d37af990ee6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.087514] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 908.087514] env[65788]: value = "task-4662627" [ 908.087514] env[65788]: _type = "Task" [ 908.087514] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.097091] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662627, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.125705] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 908.157245] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 908.157673] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.158397] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 908.158397] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.158397] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 908.158397] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 908.158604] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.158775] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 908.158958] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 908.159175] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 908.159359] env[65788]: DEBUG nova.virt.hardware [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 908.160459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9988eabe-b90a-4038-a201-e60e63332f40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.170339] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.176174] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.462s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 908.178421] env[65788]: INFO nova.compute.claims [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.188576] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028014a4-2f61-4dd1-a8c1-e4cb35b63073 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.209453] env[65788]: INFO nova.scheduler.client.report [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Deleted allocations for instance 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada [ 908.236022] env[65788]: DEBUG nova.network.neutron [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 908.299626] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f834b40a-41e6-43cb-9cb1-c97339c206be tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.447s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.347087] env[65788]: DEBUG nova.compute.manager [req-bc5558ef-eda9-4223-bc19-18d32a3e332b req-08503467-96dc-43f9-a7b4-259ca9c4ea9b service nova] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Received event network-vif-deleted-1b0ec4f3-be92-493e-a8ed-6af0f0bb844c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 908.398955] env[65788]: DEBUG oslo_vmware.api [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662625, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166853} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.399248] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.399434] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.399607] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.399777] env[65788]: INFO nova.compute.manager [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Took 1.15 seconds to destroy the instance on the hypervisor. [ 908.400222] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 908.400330] env[65788]: DEBUG nova.compute.manager [-] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 908.400396] env[65788]: DEBUG nova.network.neutron [-] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 908.400695] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.401350] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 908.401555] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 908.493560] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662626, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.601269] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662627, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081121} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.602538] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.602947] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470a9bc6-ac39-4792-ab75-d84ca553b075 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.634150] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] b4f157a7-350c-4fff-8509-44426714846c/b4f157a7-350c-4fff-8509-44426714846c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.634507] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3bc72ff-c78e-474d-8d77-cf832cc93ccf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.656255] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 908.656255] env[65788]: value = "task-4662628" [ 908.656255] env[65788]: _type = "Task" [ 908.656255] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.667325] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662628, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.673680] env[65788]: INFO nova.compute.manager [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Rescuing [ 908.673966] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.674134] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.674305] env[65788]: DEBUG nova.network.neutron [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 908.677443] env[65788]: DEBUG nova.network.neutron [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Successfully updated port: 85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 908.722181] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9985588e-90d8-478f-9682-f85fb3633198 tempest-ServerAddressesTestJSON-2021451561 tempest-ServerAddressesTestJSON-2021451561-project-member] Lock "7fd30bbe-5797-4d0f-8d01-5c8a953f2ada" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.507s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.738637] env[65788]: INFO nova.compute.manager [-] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Took 1.60 seconds to deallocate network for instance. [ 908.818558] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 908.996364] env[65788]: DEBUG oslo_vmware.api [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662626, 'name': PowerOnVM_Task, 'duration_secs': 0.876865} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.996660] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.996865] env[65788]: DEBUG nova.compute.manager [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 908.997695] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9b2871-6d4f-4570-80c7-250b3edd5008 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.169381] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662628, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.181504] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.181504] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 909.181796] env[65788]: DEBUG nova.network.neutron [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 909.187648] env[65788]: WARNING neutronclient.v2_0.client [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.187648] env[65788]: WARNING openstack [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.187648] env[65788]: WARNING openstack [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.198879] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.199293] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.249054] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.271272] env[65788]: DEBUG nova.network.neutron [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 909.377098] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.378693] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.450561] env[65788]: DEBUG nova.compute.manager [req-d73c245e-74f1-4011-9189-71386e9846cf req-0e758d27-5951-4320-9e7e-1a57331fb638 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Received event network-vif-plugged-85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 909.450789] env[65788]: DEBUG oslo_concurrency.lockutils [req-d73c245e-74f1-4011-9189-71386e9846cf req-0e758d27-5951-4320-9e7e-1a57331fb638 service nova] Acquiring lock "374564df-70fa-4b89-8f38-e559245e5ebf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.451021] env[65788]: DEBUG oslo_concurrency.lockutils [req-d73c245e-74f1-4011-9189-71386e9846cf req-0e758d27-5951-4320-9e7e-1a57331fb638 service nova] Lock "374564df-70fa-4b89-8f38-e559245e5ebf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.451205] env[65788]: DEBUG oslo_concurrency.lockutils [req-d73c245e-74f1-4011-9189-71386e9846cf req-0e758d27-5951-4320-9e7e-1a57331fb638 service nova] Lock "374564df-70fa-4b89-8f38-e559245e5ebf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.451370] env[65788]: DEBUG nova.compute.manager [req-d73c245e-74f1-4011-9189-71386e9846cf req-0e758d27-5951-4320-9e7e-1a57331fb638 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] No waiting events found dispatching network-vif-plugged-85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 909.451536] env[65788]: WARNING nova.compute.manager [req-d73c245e-74f1-4011-9189-71386e9846cf req-0e758d27-5951-4320-9e7e-1a57331fb638 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Received unexpected event network-vif-plugged-85fbe51b-2508-45d1-a3e3-b6c38fac2060 for instance with vm_state building and task_state spawning. [ 909.464797] env[65788]: WARNING openstack [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.465419] env[65788]: WARNING openstack [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.516135] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.566849] env[65788]: WARNING neutronclient.v2_0.client [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.567541] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.567883] env[65788]: WARNING openstack [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.595637] env[65788]: WARNING neutronclient.v2_0.client [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 909.596339] env[65788]: WARNING openstack [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 909.596703] env[65788]: WARNING openstack [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 909.646442] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64f3541-5e44-486d-80c7-e309433122db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.655779] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4aa3140-0bd2-4049-a770-9dc0164cab14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.675382] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662628, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.702786] env[65788]: DEBUG nova.network.neutron [-] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 909.709814] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed70e02-8919-49cd-a7cf-61e78bdcccf8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.719113] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b70279-0732-4457-9084-9815ca49a6fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.725959] env[65788]: DEBUG nova.network.neutron [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updating instance_info_cache with network_info: [{"id": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "address": "fa:16:3e:b6:76:23", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85fbe51b-25", "ovs_interfaceid": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 909.739697] env[65788]: DEBUG nova.compute.provider_tree [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.790472] env[65788]: DEBUG nova.network.neutron [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 910.066276] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "7c88e3ef-82d2-46ac-a350-999de8091c01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.066485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.066584] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "7c88e3ef-82d2-46ac-a350-999de8091c01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.066831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.067076] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.069145] env[65788]: INFO nova.compute.manager [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Terminating instance [ 910.174033] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662628, 'name': ReconfigVM_Task, 'duration_secs': 1.496426} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.174375] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Reconfigured VM instance instance-00000046 to attach disk [datastore2] b4f157a7-350c-4fff-8509-44426714846c/b4f157a7-350c-4fff-8509-44426714846c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.175208] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76fb4acd-9e76-4d46-be81-9e62a8f1bdb3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.183952] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 910.183952] env[65788]: value = "task-4662630" [ 910.183952] env[65788]: _type = "Task" [ 910.183952] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.195312] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662630, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.206036] env[65788]: INFO nova.compute.manager [-] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Took 1.81 seconds to deallocate network for instance. [ 910.231806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 910.232060] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Instance network_info: |[{"id": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "address": "fa:16:3e:b6:76:23", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85fbe51b-25", "ovs_interfaceid": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 910.232620] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:76:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604056d6-6dd6-47fa-9eaa-6863a3a7c488', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '85fbe51b-2508-45d1-a3e3-b6c38fac2060', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.240355] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Creating folder: Project (22a3db3230244ef5af1227df6f4a6f12). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.240813] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd440e95-57a2-4d74-bd63-005e4c0b64ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.243366] env[65788]: DEBUG nova.scheduler.client.report [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 910.263156] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Created folder: Project (22a3db3230244ef5af1227df6f4a6f12) in parent group-v910111. [ 910.264118] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Creating folder: Instances. Parent ref: group-v910326. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.264118] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-103a2be6-e996-438b-86df-46d6ddb82f98 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.280053] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Created folder: Instances in parent group-v910326. [ 910.280262] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 910.280575] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.280872] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b955bb74-8d8b-4e7f-a341-33606297fe05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.301832] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 910.312335] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.312335] env[65788]: value = "task-4662633" [ 910.312335] env[65788]: _type = "Task" [ 910.312335] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.324262] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662633, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.576021] env[65788]: DEBUG nova.compute.manager [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 910.576021] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.576021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1bb53c-6457-48ac-b725-ab168d7f091c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.587763] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.588081] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d72ccec4-bbc1-463a-9400-9fdde72fe825 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.597245] env[65788]: DEBUG oslo_vmware.api [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 910.597245] env[65788]: value = "task-4662634" [ 910.597245] env[65788]: _type = "Task" [ 910.597245] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.607060] env[65788]: DEBUG oslo_vmware.api [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.697245] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662630, 'name': Rename_Task, 'duration_secs': 0.168695} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.697686] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.697686] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83755d88-082c-4b2e-9dad-ed38dd18ac13 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.705896] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 910.705896] env[65788]: value = "task-4662635" [ 910.705896] env[65788]: _type = "Task" [ 910.705896] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.716150] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.716899] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662635, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.748729] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.749442] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 910.755319] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.572s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.757754] env[65788]: INFO nova.compute.claims [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.825595] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662633, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.109955] env[65788]: DEBUG oslo_vmware.api [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662634, 'name': PowerOffVM_Task, 'duration_secs': 0.247857} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.110314] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.110511] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.111079] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e9de5f0-c990-4fef-b297-0ec3ce3fcafc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.186346] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 911.186346] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 911.186346] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleting the datastore file [datastore1] 7c88e3ef-82d2-46ac-a350-999de8091c01 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.186553] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3eb7dc1b-52a0-410f-9b08-d933b2cc72b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.194847] env[65788]: DEBUG oslo_vmware.api [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 911.194847] env[65788]: value = "task-4662637" [ 911.194847] env[65788]: _type = "Task" [ 911.194847] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.205142] env[65788]: DEBUG oslo_vmware.api [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.216535] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662635, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.265418] env[65788]: DEBUG nova.compute.utils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 911.269262] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 911.269262] env[65788]: DEBUG nova.network.neutron [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 911.269570] env[65788]: WARNING neutronclient.v2_0.client [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.269873] env[65788]: WARNING neutronclient.v2_0.client [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.270572] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.270977] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.331283] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662633, 'name': CreateVM_Task, 'duration_secs': 0.541878} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.331445] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.332075] env[65788]: WARNING neutronclient.v2_0.client [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.332611] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.332887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.333280] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 911.333640] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6f07482-abaf-4ccc-afbf-8aeecdd8a7cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.345068] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 911.345068] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52290644-7dd7-fa43-1523-f1813889831d" [ 911.345068] env[65788]: _type = "Task" [ 911.345068] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.365135] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52290644-7dd7-fa43-1523-f1813889831d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.459089] env[65788]: DEBUG nova.policy [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2625758e73c64384982cb820ea055cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b70d3c8627449eaa6372ebe3bd90233', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 911.499255] env[65788]: DEBUG nova.compute.manager [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Received event network-changed-85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 911.499540] env[65788]: DEBUG nova.compute.manager [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Refreshing instance network info cache due to event network-changed-85fbe51b-2508-45d1-a3e3-b6c38fac2060. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 911.499771] env[65788]: DEBUG oslo_concurrency.lockutils [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Acquiring lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.499913] env[65788]: DEBUG oslo_concurrency.lockutils [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Acquired lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.500166] env[65788]: DEBUG nova.network.neutron [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Refreshing network info cache for port 85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 911.707695] env[65788]: DEBUG oslo_vmware.api [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274945} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.714368] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.715036] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.715036] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.715036] env[65788]: INFO nova.compute.manager [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Took 1.14 seconds to destroy the instance on the hypervisor. [ 911.715325] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 911.716853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c47d5d6-37e6-4142-8985-9ee35bbcb992 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.719677] env[65788]: DEBUG nova.compute.manager [-] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 911.719820] env[65788]: DEBUG nova.network.neutron [-] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 911.720050] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 911.720645] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 911.720960] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 911.737346] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817de5a6-5cd8-4154-b21d-2e0bfda49b09 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.740792] env[65788]: DEBUG oslo_vmware.api [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662635, 'name': PowerOnVM_Task, 'duration_secs': 0.519327} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.741899] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 911.742116] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910325', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'name': 'volume-249d7aab-deda-4273-9967-6fde28af77b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca', 'attached_at': '', 'detached_at': '', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'serial': '249d7aab-deda-4273-9967-6fde28af77b4'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 911.742538] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.742676] env[65788]: INFO nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Took 8.34 seconds to spawn the instance on the hypervisor. [ 911.742848] env[65788]: DEBUG nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 911.743619] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61fb12b-42ea-401f-86a7-af1e178278c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.747056] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43040d0-3615-4f02-82a8-b7aef77735a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.782314] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 911.799364] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ab4d04-1c6a-4410-9fbe-c2667c64ff00 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.804566] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919ff162-a25f-4b58-b81a-7cd956b4d2e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.829150] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59caaeb9-fe6f-4dd3-a792-b0ec78931f9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.841113] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] volume-249d7aab-deda-4273-9967-6fde28af77b4/volume-249d7aab-deda-4273-9967-6fde28af77b4.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.843243] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0373fc83-ac49-4f37-984a-d3049f71ec0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.870640] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 911.871189] env[65788]: DEBUG nova.compute.provider_tree [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.874783] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-529175a3-7790-4105-b612-5bead3e4476c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.876633] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 911.876633] env[65788]: value = "task-4662638" [ 911.876633] env[65788]: _type = "Task" [ 911.876633] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.884837] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52290644-7dd7-fa43-1523-f1813889831d, 'name': SearchDatastore_Task, 'duration_secs': 0.02344} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.886534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 911.886807] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.887013] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.887168] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 911.887356] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.887991] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 911.887991] env[65788]: value = "task-4662639" [ 911.887991] env[65788]: _type = "Task" [ 911.887991] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.888210] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ded497a-ffa6-4e01-894e-51a8cea459ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.897236] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662638, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.905758] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.907540] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.907735] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.908506] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aa91893-a737-4db4-a2e8-708dd8535881 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.914834] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 911.914834] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e5d8da-22f2-7f32-905a-c0dbec402451" [ 911.914834] env[65788]: _type = "Task" [ 911.914834] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.924308] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e5d8da-22f2-7f32-905a-c0dbec402451, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.005074] env[65788]: WARNING neutronclient.v2_0.client [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.005845] env[65788]: WARNING openstack [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 912.006219] env[65788]: WARNING openstack [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 912.117368] env[65788]: DEBUG nova.network.neutron [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Successfully created port: e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 912.318857] env[65788]: INFO nova.compute.manager [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Took 39.55 seconds to build instance. [ 912.320904] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 912.375028] env[65788]: DEBUG nova.scheduler.client.report [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 912.391968] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662638, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.404109] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662639, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.427773] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e5d8da-22f2-7f32-905a-c0dbec402451, 'name': SearchDatastore_Task, 'duration_secs': 0.013284} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.432542] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cd1f74d-e65a-47d0-8e32-4d9d55ca84d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.438890] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 912.438890] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f3ddf-04ed-9ebe-4a51-49085f3937b9" [ 912.438890] env[65788]: _type = "Task" [ 912.438890] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.448878] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f3ddf-04ed-9ebe-4a51-49085f3937b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.805693] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 912.824753] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7c56727-d4a0-4734-945b-acbbe0bd6efe tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "b4f157a7-350c-4fff-8509-44426714846c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.618s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.836688] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 912.836949] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 912.837489] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 912.837711] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 912.837855] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 912.838053] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 912.838280] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.838439] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 912.838605] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 912.838769] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 912.838942] env[65788]: DEBUG nova.virt.hardware [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 912.839871] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b1492e-0b7e-4a5c-938f-a3c56fd4043e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.853808] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab32c5e-44c0-4bcf-8a64-a1b8516a6ee9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.880614] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.126s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.881202] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 912.884925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.542s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.885145] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.887969] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.239s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.888214] env[65788]: DEBUG nova.objects.instance [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lazy-loading 'resources' on Instance uuid d042a386-0d16-4e2d-a8e7-7641c07bcb35 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.909303] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662639, 'name': PowerOffVM_Task, 'duration_secs': 0.590517} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.909722] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662638, 'name': ReconfigVM_Task, 'duration_secs': 0.619882} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.909993] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 912.911463] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Reconfigured VM instance instance-00000045 to attach disk [datastore2] volume-249d7aab-deda-4273-9967-6fde28af77b4/volume-249d7aab-deda-4273-9967-6fde28af77b4.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.918903] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d7c91b-0aa0-42ad-b7b6-d3920db74687 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.924296] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa29e581-e0aa-4dca-9586-aef941b32242 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.946992] env[65788]: INFO nova.scheduler.client.report [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocations for instance f75000e6-f3bf-4951-a541-3aa185339054 [ 912.983933] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb391a3-2cdc-451e-ae3e-d80238f0979c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.986900] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 912.986900] env[65788]: value = "task-4662640" [ 912.986900] env[65788]: _type = "Task" [ 912.986900] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.991643] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f3ddf-04ed-9ebe-4a51-49085f3937b9, 'name': SearchDatastore_Task, 'duration_secs': 0.019308} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.997242] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 912.997242] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 374564df-70fa-4b89-8f38-e559245e5ebf/374564df-70fa-4b89-8f38-e559245e5ebf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.999880] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c697c13e-6619-44a6-b16d-fe0df250de22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.012379] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662640, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.014509] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 913.014509] env[65788]: value = "task-4662641" [ 913.014509] env[65788]: _type = "Task" [ 913.014509] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.023899] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.051924] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.052743] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea6233fa-a639-438f-84a1-71f25b260f42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.063199] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 913.063199] env[65788]: value = "task-4662642" [ 913.063199] env[65788]: _type = "Task" [ 913.063199] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.075016] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 913.075330] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.075558] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.075737] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 913.075935] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.076618] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eeb871a1-7260-4a1c-8b35-9e9b3bef741c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.089522] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.089725] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.090626] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c6ad131-2bf2-4fb0-a516-06b9300d89db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.098112] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 913.098112] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522343d8-ef1a-a145-7c0b-8a9c6d1c0208" [ 913.098112] env[65788]: _type = "Task" [ 913.098112] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.108750] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522343d8-ef1a-a145-7c0b-8a9c6d1c0208, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.388137] env[65788]: DEBUG nova.compute.utils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 913.390913] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 913.392541] env[65788]: DEBUG nova.network.neutron [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 913.393156] env[65788]: WARNING neutronclient.v2_0.client [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.393502] env[65788]: WARNING neutronclient.v2_0.client [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 913.396992] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 913.397436] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 913.490726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a7f8282b-4b8c-4942-a900-98cc3ff8c414 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "f75000e6-f3bf-4951-a541-3aa185339054" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.624s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 913.506331] env[65788]: DEBUG oslo_vmware.api [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662640, 'name': ReconfigVM_Task, 'duration_secs': 0.20779} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.506687] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910325', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'name': 'volume-249d7aab-deda-4273-9967-6fde28af77b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca', 'attached_at': '', 'detached_at': '', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'serial': '249d7aab-deda-4273-9967-6fde28af77b4'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 913.529321] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662641, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.609578] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522343d8-ef1a-a145-7c0b-8a9c6d1c0208, 'name': SearchDatastore_Task, 'duration_secs': 0.022434} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.614034] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4976cb5d-818a-40f1-8e58-973809582fed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.620954] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 913.620954] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287d2d6-e6f1-2d72-29c9-f259fa3da91f" [ 913.620954] env[65788]: _type = "Task" [ 913.620954] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.632189] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287d2d6-e6f1-2d72-29c9-f259fa3da91f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.819489] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425679dc-0477-43bb-ab07-63b826f8fde9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.829033] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b66307-33fa-4d5b-92bb-0763c9e1351c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.861846] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81a75445-9599-46be-8745-6b5729825bb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.871985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20c6f4d-781c-411d-8ed9-fe52c4ff1475 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.887768] env[65788]: DEBUG nova.compute.provider_tree [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.909218] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 914.026603] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.986031} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.027107] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 374564df-70fa-4b89-8f38-e559245e5ebf/374564df-70fa-4b89-8f38-e559245e5ebf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 914.027444] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.027893] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3aab356-f201-45ee-81e5-e00e2f9c93c6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.037545] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 914.037545] env[65788]: value = "task-4662643" [ 914.037545] env[65788]: _type = "Task" [ 914.037545] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.048109] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.128765] env[65788]: DEBUG nova.network.neutron [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Successfully updated port: e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 914.137519] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287d2d6-e6f1-2d72-29c9-f259fa3da91f, 'name': SearchDatastore_Task, 'duration_secs': 0.079937} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.137833] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 914.138045] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. {{(pid=65788) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 914.138375] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e7c70d6-b358-4e9c-a0fe-1bb4bec7ffa7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.148230] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 914.148230] env[65788]: value = "task-4662644" [ 914.148230] env[65788]: _type = "Task" [ 914.148230] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.160437] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.253450] env[65788]: DEBUG nova.policy [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b0c1ce04a704a3eaf0a0d4dd09f7e08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9674e2a5c86b48db8c865a50331ab846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 914.290502] env[65788]: WARNING openstack [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 914.290502] env[65788]: WARNING openstack [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 914.391237] env[65788]: DEBUG nova.scheduler.client.report [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 914.552130] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076663} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.552130] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.552130] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67c8c80-5a14-491d-80b9-46b9ba3e43e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.575035] env[65788]: DEBUG nova.objects.instance [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'flavor' on Instance uuid 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.589480] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 374564df-70fa-4b89-8f38-e559245e5ebf/374564df-70fa-4b89-8f38-e559245e5ebf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.590674] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6155f57-34c9-4737-af37-80d1de14eef0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.613436] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 914.613436] env[65788]: value = "task-4662645" [ 914.613436] env[65788]: _type = "Task" [ 914.613436] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.621143] env[65788]: DEBUG nova.network.neutron [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Successfully created port: cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 914.627880] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662645, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.632875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.632875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.632875] env[65788]: DEBUG nova.network.neutron [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 914.659409] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.688988] env[65788]: DEBUG nova.network.neutron [-] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 914.897540] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.009s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.900139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.663s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.900786] env[65788]: DEBUG nova.objects.instance [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lazy-loading 'resources' on Instance uuid 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 914.920877] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 914.927651] env[65788]: INFO nova.scheduler.client.report [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleted allocations for instance d042a386-0d16-4e2d-a8e7-7641c07bcb35 [ 914.954685] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 914.954977] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 914.955267] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 914.955399] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 914.955579] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 914.955713] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 914.955957] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.956184] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 914.956493] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 914.956558] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 914.956710] env[65788]: DEBUG nova.virt.hardware [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 914.957677] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e79720-0965-4f17-8169-56ae4931dd4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.967201] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cbfc97-efa7-4eee-b42b-8e10e6f65220 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.093343] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c8ee8c04-93ea-4cd5-ad4e-782299d90eb4 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.512s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.128922] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662645, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.135913] env[65788]: WARNING neutronclient.v2_0.client [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.136606] env[65788]: WARNING openstack [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.137087] env[65788]: WARNING openstack [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.147028] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.147234] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.165643] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.192825] env[65788]: INFO nova.compute.manager [-] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Took 3.47 seconds to deallocate network for instance. [ 915.236818] env[65788]: DEBUG nova.network.neutron [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 915.436273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0162f957-b2db-45da-8296-41433bd74946 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "d042a386-0d16-4e2d-a8e7-7641c07bcb35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.500s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.466529] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.466529] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 915.521705] env[65788]: DEBUG nova.network.neutron [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updated VIF entry in instance network info cache for port 85fbe51b-2508-45d1-a3e3-b6c38fac2060. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 915.523017] env[65788]: DEBUG nova.network.neutron [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updating instance_info_cache with network_info: [{"id": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "address": "fa:16:3e:b6:76:23", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85fbe51b-25", "ovs_interfaceid": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 915.634019] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662645, 'name': ReconfigVM_Task, 'duration_secs': 0.863267} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.634411] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 374564df-70fa-4b89-8f38-e559245e5ebf/374564df-70fa-4b89-8f38-e559245e5ebf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.635035] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-284fb307-b66e-4d89-abc0-b3b363726770 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.649023] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 915.649023] env[65788]: value = "task-4662646" [ 915.649023] env[65788]: _type = "Task" [ 915.649023] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.660960] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662646, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.680988] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662644, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.702899] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.861488] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5f2aed-7c4c-463c-a1b6-25655bbaac0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.870785] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f49461-b86a-45b9-b7b3-61765cef9593 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.906874] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4395ae01-e816-454a-9236-aa84cc5154c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.915619] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063ac658-1fd3-4c73-8ac7-02266b5ad1a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.931320] env[65788]: DEBUG nova.compute.provider_tree [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.970079] env[65788]: WARNING neutronclient.v2_0.client [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 915.970879] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 915.971307] env[65788]: WARNING openstack [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 916.025998] env[65788]: DEBUG oslo_concurrency.lockutils [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] Releasing lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 916.026312] env[65788]: DEBUG nova.compute.manager [req-e96455c8-05ba-43dd-96d0-f7209211564b req-a91cd78b-fb62-45c9-be9c-b42dcee3e824 service nova] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Received event network-vif-deleted-680d08e3-6217-4e78-82fa-ea979448f95c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 916.163040] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662646, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.170110] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662644, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.69778} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.170407] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. [ 916.171228] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5080a3db-0e47-4d2d-8e0e-42a356fc4beb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.202928] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.203617] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69e7ff7c-127a-4dcc-8a4f-061520709c9b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.226352] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 916.226352] env[65788]: value = "task-4662647" [ 916.226352] env[65788]: _type = "Task" [ 916.226352] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.237873] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662647, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.327806] env[65788]: DEBUG nova.network.neutron [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Successfully updated port: cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 916.403054] env[65788]: DEBUG nova.network.neutron [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 916.415363] env[65788]: DEBUG nova.compute.manager [req-b246e0ef-23d8-4f7b-957d-320fbca926fb req-68dc7054-8d13-4318-8593-0a7ff16d4e75 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Received event network-vif-plugged-e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 916.415662] env[65788]: DEBUG oslo_concurrency.lockutils [req-b246e0ef-23d8-4f7b-957d-320fbca926fb req-68dc7054-8d13-4318-8593-0a7ff16d4e75 service nova] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.416069] env[65788]: DEBUG oslo_concurrency.lockutils [req-b246e0ef-23d8-4f7b-957d-320fbca926fb req-68dc7054-8d13-4318-8593-0a7ff16d4e75 service nova] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.416244] env[65788]: DEBUG oslo_concurrency.lockutils [req-b246e0ef-23d8-4f7b-957d-320fbca926fb req-68dc7054-8d13-4318-8593-0a7ff16d4e75 service nova] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.416474] env[65788]: DEBUG nova.compute.manager [req-b246e0ef-23d8-4f7b-957d-320fbca926fb req-68dc7054-8d13-4318-8593-0a7ff16d4e75 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] No waiting events found dispatching network-vif-plugged-e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 916.416700] env[65788]: WARNING nova.compute.manager [req-b246e0ef-23d8-4f7b-957d-320fbca926fb req-68dc7054-8d13-4318-8593-0a7ff16d4e75 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Received unexpected event network-vif-plugged-e98f1b15-0a05-4be4-b950-dd7e6749eb8f for instance with vm_state building and task_state spawning. [ 916.434584] env[65788]: DEBUG nova.scheduler.client.report [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 916.445068] env[65788]: DEBUG nova.compute.manager [req-52af4b79-19ba-4b92-b417-ee9996902ab6 req-374cabb3-77c3-49e4-9edf-e3b7102ebd90 service nova] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Received event network-vif-deleted-08be336c-c7bd-448b-b11d-931b7e9788c9 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 916.659820] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662646, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.743883] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662647, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.831738] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.831866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.832015] env[65788]: DEBUG nova.network.neutron [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 916.905844] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 916.906274] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Instance network_info: |[{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 916.906925] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:9d:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e98f1b15-0a05-4be4-b950-dd7e6749eb8f', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.916050] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 916.917116] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 916.917451] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb66fd49-2d03-42bf-baff-e36b31e5b20b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.941242] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.041s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.944441] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.944441] env[65788]: value = "task-4662648" [ 916.944441] env[65788]: _type = "Task" [ 916.944441] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.945204] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.786s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.946293] env[65788]: DEBUG nova.objects.instance [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lazy-loading 'resources' on Instance uuid ed55713e-8bf1-4960-8309-47498b163d02 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.961386] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.003578] env[65788]: INFO nova.scheduler.client.report [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Deleted allocations for instance 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468 [ 917.026980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.027348] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.027597] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.027835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.027996] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.032032] env[65788]: INFO nova.compute.manager [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Terminating instance [ 917.112545] env[65788]: INFO nova.compute.manager [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Rescuing [ 917.112820] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.112973] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 917.113157] env[65788]: DEBUG nova.network.neutron [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 917.161920] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662646, 'name': Rename_Task, 'duration_secs': 1.152039} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.162561] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.163014] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d8c0ffe-c158-4507-8d39-663197bbfe2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.172078] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 917.172078] env[65788]: value = "task-4662649" [ 917.172078] env[65788]: _type = "Task" [ 917.172078] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.183979] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.241466] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662647, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.335764] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.336190] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.455797] env[65788]: DEBUG nova.network.neutron [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 917.465268] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.482626] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.483098] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.515624] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52481707-4957-4899-b9e8-e03c0df0bd1e tempest-AttachInterfacesV270Test-88998432 tempest-AttachInterfacesV270Test-88998432-project-member] Lock "34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.470s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.536163] env[65788]: DEBUG nova.compute.manager [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 917.536359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.536741] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2356adce-ccc1-43f3-bd71-ffa9c7fcb856 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.548316] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 917.548316] env[65788]: value = "task-4662650" [ 917.548316] env[65788]: _type = "Task" [ 917.548316] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.562803] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.617486] env[65788]: WARNING neutronclient.v2_0.client [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.618526] env[65788]: WARNING openstack [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.619087] env[65788]: WARNING openstack [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.645127] env[65788]: WARNING neutronclient.v2_0.client [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 917.645127] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.645127] env[65788]: WARNING openstack [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.686351] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.740456] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662647, 'name': ReconfigVM_Task, 'duration_secs': 1.072179} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.745660] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.749271] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0023aea6-0fc5-4ae8-943e-5ff6e6145ed2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.782935] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f463b0b5-0941-4b67-b013-68f73ce3467e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.801779] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 917.801779] env[65788]: value = "task-4662651" [ 917.801779] env[65788]: _type = "Task" [ 917.801779] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.816419] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.869042] env[65788]: DEBUG nova.network.neutron [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance_info_cache with network_info: [{"id": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "address": "fa:16:3e:0a:42:08", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2184dd-c8", "ovs_interfaceid": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 917.907644] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec2b38f-4bec-4d88-8758-cdb1c6d903b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.917506] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b50f9c-c50f-4ca6-a118-eacaef012112 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.924213] env[65788]: WARNING openstack [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 917.924831] env[65788]: WARNING openstack [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 917.969227] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eee7c6b-786f-4744-bd64-6403363777a1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.989135] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db0b7ad-63dc-451d-8128-7d66fa00033d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.993354] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.007244] env[65788]: DEBUG nova.compute.provider_tree [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.059782] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662650, 'name': PowerOffVM_Task, 'duration_secs': 0.384598} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.060881] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.060881] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 918.060881] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910325', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'name': 'volume-249d7aab-deda-4273-9967-6fde28af77b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca', 'attached_at': '', 'detached_at': '', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'serial': '249d7aab-deda-4273-9967-6fde28af77b4'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 918.061346] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049202db-8e5b-474d-8f1a-7d8bf531dfad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.085010] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1ca169-6ad4-4d72-ac3f-8ab70b0de3aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.093775] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2570d415-0637-4b38-947d-22faab0d9cd0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.122746] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2813ba97-b1e7-46d1-b362-445f2b610437 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.140593] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The volume has not been displaced from its original location: [datastore2] volume-249d7aab-deda-4273-9967-6fde28af77b4/volume-249d7aab-deda-4273-9967-6fde28af77b4.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 918.146054] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 918.146644] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66ead9d2-ee0d-4ff1-9777-58582c27bb5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.172669] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 918.172669] env[65788]: value = "task-4662652" [ 918.172669] env[65788]: _type = "Task" [ 918.172669] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.187692] env[65788]: WARNING neutronclient.v2_0.client [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 918.188795] env[65788]: WARNING openstack [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 918.189391] env[65788]: WARNING openstack [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 918.205694] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.209319] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.313296] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.331658] env[65788]: DEBUG nova.network.neutron [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 918.372793] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 918.373182] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Instance network_info: |[{"id": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "address": "fa:16:3e:0a:42:08", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2184dd-c8", "ovs_interfaceid": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 918.374139] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0a:42:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd2184dd-c826-412e-a259-8e9ebb22c2e5', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.382274] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 918.382943] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.384113] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49e554ce-a566-4ca9-8656-2608e32f782c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.407826] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.407826] env[65788]: value = "task-4662653" [ 918.407826] env[65788]: _type = "Task" [ 918.407826] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.417262] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662653, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.473655] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.510878] env[65788]: DEBUG nova.scheduler.client.report [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 918.688648] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.695323] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.819095] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.837583] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 918.919457] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662653, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.978273] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.020030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.020030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.020030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.074s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.023670] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.038s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.024270] env[65788]: DEBUG nova.objects.instance [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'resources' on Instance uuid 5e27dbf0-44d2-46a4-87b2-c209afb9559e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.061688] env[65788]: INFO nova.scheduler.client.report [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Deleted allocations for instance ed55713e-8bf1-4960-8309-47498b163d02 [ 919.193731] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.194012] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.316962] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.320086] env[65788]: DEBUG nova.compute.manager [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Received event network-changed-e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 919.320086] env[65788]: DEBUG nova.compute.manager [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Refreshing instance network info cache due to event network-changed-e98f1b15-0a05-4be4-b950-dd7e6749eb8f. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 919.320509] env[65788]: DEBUG oslo_concurrency.lockutils [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.320679] env[65788]: DEBUG oslo_concurrency.lockutils [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.320847] env[65788]: DEBUG nova.network.neutron [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Refreshing network info cache for port e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 919.346143] env[65788]: DEBUG nova.compute.manager [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Received event network-vif-plugged-cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 919.346399] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Acquiring lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.346960] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.347198] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.347710] env[65788]: DEBUG nova.compute.manager [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] No waiting events found dispatching network-vif-plugged-cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 919.347710] env[65788]: WARNING nova.compute.manager [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Received unexpected event network-vif-plugged-cd2184dd-c826-412e-a259-8e9ebb22c2e5 for instance with vm_state building and task_state spawning. [ 919.347710] env[65788]: DEBUG nova.compute.manager [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Received event network-changed-cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 919.347848] env[65788]: DEBUG nova.compute.manager [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Refreshing instance network info cache due to event network-changed-cd2184dd-c826-412e-a259-8e9ebb22c2e5. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 919.348011] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Acquiring lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.348167] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Acquired lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.348324] env[65788]: DEBUG nova.network.neutron [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Refreshing network info cache for port cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 919.420924] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662653, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.479943] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.530119] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 919.571658] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3c9283a-8b79-44a8-bf6f-36f67b2ea26d tempest-MigrationsAdminTest-1412758904 tempest-MigrationsAdminTest-1412758904-project-member] Lock "ed55713e-8bf1-4960-8309-47498b163d02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.939s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 919.635479] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.635716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 919.692940] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.699915] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.821550] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.827070] env[65788]: WARNING neutronclient.v2_0.client [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.827655] env[65788]: WARNING openstack [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.828020] env[65788]: WARNING openstack [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.852868] env[65788]: WARNING neutronclient.v2_0.client [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 919.853117] env[65788]: WARNING openstack [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 919.853472] env[65788]: WARNING openstack [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 919.919934] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662653, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.988047] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.990528] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a0c309-df74-4129-824d-fe8c6de96496 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.998818] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba54c23-4d3b-4b4c-8593-45c27fc4f41c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.035018] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d43070-9fc9-4106-8076-4278c7c0f4bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.044190] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c22e4d0-a0a0-47be-b15e-4499a7487b70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.065950] env[65788]: DEBUG nova.compute.provider_tree [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.070676] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.109386] env[65788]: WARNING openstack [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.110071] env[65788]: WARNING openstack [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.138303] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 920.169034] env[65788]: WARNING openstack [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.169393] env[65788]: WARNING openstack [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.178075] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "e0f7c849-315c-4247-a840-d388d48746b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.178334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "e0f7c849-315c-4247-a840-d388d48746b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.194940] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.198604] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.316799] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.322527] env[65788]: WARNING neutronclient.v2_0.client [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.323215] env[65788]: WARNING openstack [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.323575] env[65788]: WARNING openstack [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.386941] env[65788]: WARNING neutronclient.v2_0.client [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.388820] env[65788]: WARNING openstack [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 920.388820] env[65788]: WARNING openstack [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 920.397652] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.397880] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de3fed2a-087b-4cad-8b85-e9744fa781b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.410391] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 920.410391] env[65788]: value = "task-4662654" [ 920.410391] env[65788]: _type = "Task" [ 920.410391] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.423053] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.426486] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662653, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.480182] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.561792] env[65788]: DEBUG nova.network.neutron [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updated VIF entry in instance network info cache for port e98f1b15-0a05-4be4-b950-dd7e6749eb8f. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 920.562446] env[65788]: DEBUG nova.network.neutron [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 920.572358] env[65788]: DEBUG nova.scheduler.client.report [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 920.618621] env[65788]: DEBUG nova.network.neutron [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updated VIF entry in instance network info cache for port cd2184dd-c826-412e-a259-8e9ebb22c2e5. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 920.619012] env[65788]: DEBUG nova.network.neutron [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance_info_cache with network_info: [{"id": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "address": "fa:16:3e:0a:42:08", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2184dd-c8", "ovs_interfaceid": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 920.670591] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.689301] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 920.701674] env[65788]: DEBUG oslo_vmware.api [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662649, 'name': PowerOnVM_Task, 'duration_secs': 3.284869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.705960] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.706333] env[65788]: INFO nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Took 12.58 seconds to spawn the instance on the hypervisor. [ 920.706570] env[65788]: DEBUG nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 920.707494] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.708101] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0db0392-6216-4366-bd89-2463c8027185 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.818958] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662651, 'name': ReconfigVM_Task, 'duration_secs': 2.742019} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.819338] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.819667] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-989bd4b0-b86d-4f1f-899f-6b5e2dd048de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.830269] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 920.830269] env[65788]: value = "task-4662655" [ 920.830269] env[65788]: _type = "Task" [ 920.830269] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.840374] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.931057] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662654, 'name': PowerOffVM_Task, 'duration_secs': 0.299387} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.931315] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662653, 'name': CreateVM_Task, 'duration_secs': 2.436364} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.931544] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 920.931795] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.932762] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91092155-8bce-4f61-832a-83fea6141cce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.935808] env[65788]: WARNING neutronclient.v2_0.client [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.936200] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.936340] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.936670] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 920.936957] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0632875-a7b0-4017-bb8c-6a7e396279c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.943828] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 920.943828] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e87bf3-bbb5-c740-fca6-044650688af2" [ 920.943828] env[65788]: _type = "Task" [ 920.943828] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.966134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7880989-2f38-4670-8d6a-859840b96ca4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.977653] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e87bf3-bbb5-c740-fca6-044650688af2, 'name': SearchDatastore_Task, 'duration_secs': 0.009749} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.984646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 920.984646] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.984646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.984646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.984781] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.989218] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07bdd512-3c27-4f4b-936b-d05638fcc221 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.991328] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662648, 'name': CreateVM_Task, 'duration_secs': 3.625267} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.992025] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.992512] env[65788]: WARNING neutronclient.v2_0.client [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 920.992798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.992941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.993468] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 920.993574] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-180e89d4-f97f-4b84-b852-f84b6c2590e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.999085] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 920.999085] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f68bb1-acc2-95fa-c455-8b2357cef0ca" [ 920.999085] env[65788]: _type = "Task" [ 920.999085] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.001109] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.001412] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.005635] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f0d9dc4-787a-4d40-bc87-4557f4b545fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.012599] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 921.012599] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c530b0-97a4-1f59-ab79-9fed92c487e6" [ 921.012599] env[65788]: _type = "Task" [ 921.012599] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.016755] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f68bb1-acc2-95fa-c455-8b2357cef0ca, 'name': SearchDatastore_Task, 'duration_secs': 0.010137} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.022883] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.023184] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.023426] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.030342] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c530b0-97a4-1f59-ab79-9fed92c487e6, 'name': SearchDatastore_Task, 'duration_secs': 0.010265} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.032370] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.033272] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ad5458d-392f-4eb3-a038-c4e0a6ac229f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.034973] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8078d4bf-16b7-4e90-ba1b-97e623feadd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.041078] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 921.041078] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52efee5b-24fb-f5e7-4d35-e1866f47d269" [ 921.041078] env[65788]: _type = "Task" [ 921.041078] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.046210] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 921.046210] env[65788]: value = "task-4662656" [ 921.046210] env[65788]: _type = "Task" [ 921.046210] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.052612] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52efee5b-24fb-f5e7-4d35-e1866f47d269, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.058496] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 921.058708] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.058927] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.067673] env[65788]: DEBUG oslo_concurrency.lockutils [req-51d4d4f3-0404-4eea-9c96-5d8e327c41cf req-9b521074-ea40-4ab7-90bc-1df023f89d39 service nova] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.077651] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.055s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.080009] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.597s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.080284] env[65788]: DEBUG nova.objects.instance [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lazy-loading 'resources' on Instance uuid 2c24bc30-d413-4714-81c2-b657cafe94bd {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.108844] env[65788]: INFO nova.scheduler.client.report [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted allocations for instance 5e27dbf0-44d2-46a4-87b2-c209afb9559e [ 921.122035] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc2d58df-25b3-4b93-98f4-68c9247eadcf req-d98e20f2-92af-49b3-bccb-3b4a520b234e service nova] Releasing lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.190867] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662652, 'name': ReconfigVM_Task, 'duration_secs': 2.609866} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.191081] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 921.200698] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90eea04c-a5d5-48a5-aca9-09996b62bf82 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.220578] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 921.220578] env[65788]: value = "task-4662657" [ 921.220578] env[65788]: _type = "Task" [ 921.220578] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.227015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.232573] env[65788]: INFO nova.compute.manager [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Took 41.33 seconds to build instance. [ 921.237330] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662657, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.340969] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.554888] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52efee5b-24fb-f5e7-4d35-e1866f47d269, 'name': SearchDatastore_Task, 'duration_secs': 0.010712} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.555265] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.555582] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e/2cb53fe5-aa58-479e-9090-0d8509cf164e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.556025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.556244] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.556513] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75d80af9-b9a0-411b-b754-37b0074cdf1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.558750] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de038dc7-b56d-4ef6-b04e-b0d0208b79d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.566342] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 921.566342] env[65788]: value = "task-4662658" [ 921.566342] env[65788]: _type = "Task" [ 921.566342] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.571040] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.571309] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.572583] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f525d1a-e51a-4ce0-996d-8e31e896875b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.586023] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.594016] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 921.594016] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5222ada0-0491-1efd-b63d-f04330460c5e" [ 921.594016] env[65788]: _type = "Task" [ 921.594016] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.607789] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5222ada0-0491-1efd-b63d-f04330460c5e, 'name': SearchDatastore_Task, 'duration_secs': 0.013177} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.609351] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ac6bc63-a4f2-488c-b589-f61df9aba659 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.621537] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 921.621537] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522757c9-1d18-41c0-fe1c-c4104df61924" [ 921.621537] env[65788]: _type = "Task" [ 921.621537] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.622195] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e2ada1d-63ef-4ec3-9d42-3353c319c202 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "5e27dbf0-44d2-46a4-87b2-c209afb9559e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.208s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.635833] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522757c9-1d18-41c0-fe1c-c4104df61924, 'name': SearchDatastore_Task, 'duration_secs': 0.010273} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.636398] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 921.636615] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.636923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 921.637199] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.637444] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26458586-2d66-4e38-8714-0623daba95e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.641344] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ff132e6-c086-43ba-8365-83ff732e0002 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.651138] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 921.651138] env[65788]: value = "task-4662659" [ 921.651138] env[65788]: _type = "Task" [ 921.651138] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.652616] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.652837] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.657625] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-879a9389-88af-40d1-a9fe-ef640f4ba724 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.674669] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662659, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.676019] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 921.676019] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5282985a-dfb6-0fd0-237a-690bd0e6ea88" [ 921.676019] env[65788]: _type = "Task" [ 921.676019] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.692381] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5282985a-dfb6-0fd0-237a-690bd0e6ea88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.735558] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662657, 'name': ReconfigVM_Task, 'duration_secs': 0.217929} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.738452] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910325', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'name': 'volume-249d7aab-deda-4273-9967-6fde28af77b4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca', 'attached_at': '', 'detached_at': '', 'volume_id': '249d7aab-deda-4273-9967-6fde28af77b4', 'serial': '249d7aab-deda-4273-9967-6fde28af77b4'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 921.738775] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.739217] env[65788]: DEBUG oslo_concurrency.lockutils [None req-333b9acd-7dfd-43b1-afb3-1a119c572461 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "374564df-70fa-4b89-8f38-e559245e5ebf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.843s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.740755] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787901b4-ab9f-4660-bc6f-132ab2d90fb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.751601] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.752664] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-649bef9d-e614-4d20-83a4-74752015d33b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.831150] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.831420] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.831586] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleting the datastore file [datastore1] 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.836098] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8be17337-134c-4a3e-b613-02ff84dcbec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.850583] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 921.850583] env[65788]: value = "task-4662661" [ 921.850583] env[65788]: _type = "Task" [ 921.850583] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.856405] env[65788]: DEBUG oslo_vmware.api [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662655, 'name': PowerOnVM_Task, 'duration_secs': 0.803581} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.860812] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.865877] env[65788]: DEBUG nova.compute.manager [None req-ddb4746d-1733-4d6c-988f-202fe9696ecc tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 921.868636] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4308becb-7970-4813-94ef-b918f16667ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.875139] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662661, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.074824] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ff956b-39ea-418f-87f8-cfb98cee802f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.084426] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662658, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.088367] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881f8bd6-6297-45c5-98c0-fcfd26f7ebba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.128010] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44cb635-a1bf-476a-aa7a-3eb913e36780 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.136878] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6924b410-6f31-4f83-bccf-a7cb82b3b4b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.157512] env[65788]: DEBUG nova.compute.provider_tree [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.170667] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662659, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.191392] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5282985a-dfb6-0fd0-237a-690bd0e6ea88, 'name': SearchDatastore_Task, 'duration_secs': 0.024881} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.193557] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4962e768-1865-4398-9f26-5b7dab09ad85 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.202116] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 922.202116] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be2c3f-4e49-1340-8064-08c73505629c" [ 922.202116] env[65788]: _type = "Task" [ 922.202116] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.214907] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be2c3f-4e49-1340-8064-08c73505629c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.370274] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662661, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.586369] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662658, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552317} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.586369] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e/2cb53fe5-aa58-479e-9090-0d8509cf164e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.586369] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.586369] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54236040-24dc-4a33-934c-1b48c9f385d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.594424] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 922.594424] env[65788]: value = "task-4662662" [ 922.594424] env[65788]: _type = "Task" [ 922.594424] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.606143] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662662, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.663705] env[65788]: DEBUG nova.scheduler.client.report [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 922.675012] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662659, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.81181} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.675410] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.676119] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.676978] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d5bd263-614d-486a-9d68-070e1e9221d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.686316] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 922.686316] env[65788]: value = "task-4662663" [ 922.686316] env[65788]: _type = "Task" [ 922.686316] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.697950] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.715642] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be2c3f-4e49-1340-8064-08c73505629c, 'name': SearchDatastore_Task, 'duration_secs': 0.060897} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.716631] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.717186] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b4f157a7-350c-4fff-8509-44426714846c/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. {{(pid=65788) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 922.717594] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad029c10-8aed-4014-a805-dbda779ac363 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.726828] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 922.726828] env[65788]: value = "task-4662664" [ 922.726828] env[65788]: _type = "Task" [ 922.726828] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.737439] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662664, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.870430] env[65788]: DEBUG oslo_vmware.api [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662661, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.573951} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.871607] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.871607] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.871607] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.871607] env[65788]: INFO nova.compute.manager [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Took 5.34 seconds to destroy the instance on the hypervisor. [ 922.871607] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 922.871831] env[65788]: DEBUG nova.compute.manager [-] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 922.871914] env[65788]: DEBUG nova.network.neutron [-] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 922.872747] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 922.872842] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 922.874132] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 923.109965] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662662, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091537} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.110374] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.111155] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426d2339-8e0d-4412-8b18-1e439605a4b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.139159] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e/2cb53fe5-aa58-479e-9090-0d8509cf164e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.139777] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-061151b7-c64a-4ee1-a644-7613190edee6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.167093] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 923.167093] env[65788]: value = "task-4662665" [ 923.167093] env[65788]: _type = "Task" [ 923.167093] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.177363] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.097s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.185302] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.936s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 923.185755] env[65788]: DEBUG nova.objects.instance [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lazy-loading 'resources' on Instance uuid 5cef20ed-d9f0-4237-ae9d-da401b953904 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.187299] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662665, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.202873] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133739} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.204022] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.205202] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b07d37-b194-4341-89cf-46a093eab68f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.236685] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.238298] env[65788]: INFO nova.scheduler.client.report [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Deleted allocations for instance 2c24bc30-d413-4714-81c2-b657cafe94bd [ 923.241436] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c522be08-4918-4b7e-8f0e-ffd7327a869e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.270626] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662664, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51475} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.272183] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b4f157a7-350c-4fff-8509-44426714846c/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. [ 923.272688] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 923.272688] env[65788]: value = "task-4662666" [ 923.272688] env[65788]: _type = "Task" [ 923.272688] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.274082] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb03de8f-3bc2-4469-9498-70d1cc82a14d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.312064] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] b4f157a7-350c-4fff-8509-44426714846c/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.312990] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e9df70c-b8f2-42c7-8846-1f135d42490a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.327017] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662666, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.333617] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 923.333617] env[65788]: value = "task-4662667" [ 923.333617] env[65788]: _type = "Task" [ 923.333617] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.344370] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662667, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.682445] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662665, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.772920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c71fd517-4bd5-4f5e-abbb-931a44e1674b tempest-ListImageFiltersTestJSON-600591362 tempest-ListImageFiltersTestJSON-600591362-project-member] Lock "2c24bc30-d413-4714-81c2-b657cafe94bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.016s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 923.795897] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662666, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.852232] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.944968] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 924.108522] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f75323-a093-4e6c-8480-fb1cc79aaac0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.118702] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f19d37-ef56-4b63-b6e5-e5369b9e5f97 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.163519] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3bbbe6-5919-44a4-a408-30bf019f5ad8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.177731] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0067de-4aeb-4c55-a039-04d650ad0500 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.186320] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662665, 'name': ReconfigVM_Task, 'duration_secs': 0.735781} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.187572] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e/2cb53fe5-aa58-479e-9090-0d8509cf164e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.188019] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe07096d-ef93-4a9c-b2bc-3266b05e91dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.202170] env[65788]: DEBUG nova.compute.provider_tree [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.210864] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 924.210864] env[65788]: value = "task-4662668" [ 924.210864] env[65788]: _type = "Task" [ 924.210864] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.222044] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662668, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.292859] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662666, 'name': ReconfigVM_Task, 'duration_secs': 0.764012} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.293228] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.295157] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6eacb93d-b4fe-4ddc-a042-a3e1eb229023 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.306726] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 924.306726] env[65788]: value = "task-4662669" [ 924.306726] env[65788]: _type = "Task" [ 924.306726] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.318983] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662669, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.350047] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662667, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.707619] env[65788]: DEBUG nova.scheduler.client.report [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 924.727930] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662668, 'name': Rename_Task, 'duration_secs': 0.166009} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.728278] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.728562] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15521d75-a3ef-412e-b01d-a055556641d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.740671] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 924.740671] env[65788]: value = "task-4662670" [ 924.740671] env[65788]: _type = "Task" [ 924.740671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.750687] env[65788]: DEBUG nova.compute.manager [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Received event network-changed-85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 924.751800] env[65788]: DEBUG nova.compute.manager [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Refreshing instance network info cache due to event network-changed-85fbe51b-2508-45d1-a3e3-b6c38fac2060. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 924.752471] env[65788]: DEBUG oslo_concurrency.lockutils [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Acquiring lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.752649] env[65788]: DEBUG oslo_concurrency.lockutils [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Acquired lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 924.752823] env[65788]: DEBUG nova.network.neutron [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Refreshing network info cache for port 85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 924.762328] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662670, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.789050] env[65788]: DEBUG nova.network.neutron [-] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 924.817733] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662669, 'name': Rename_Task, 'duration_secs': 0.163127} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.817733] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.817733] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44042792-ab4c-4c35-ab0d-97698a8756c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.827839] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 924.827839] env[65788]: value = "task-4662671" [ 924.827839] env[65788]: _type = "Task" [ 924.827839] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.840040] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.850077] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662667, 'name': ReconfigVM_Task, 'duration_secs': 1.069888} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.850794] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Reconfigured VM instance instance-00000046 to attach disk [datastore2] b4f157a7-350c-4fff-8509-44426714846c/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.852441] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea128b9-c864-4bca-ab3c-90f4b76489b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.883084] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ad504a3-6846-4409-9114-65b328522c9a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.902463] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 924.902463] env[65788]: value = "task-4662672" [ 924.902463] env[65788]: _type = "Task" [ 924.902463] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.912646] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662672, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.221213] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.035s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.225407] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.708s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 925.225744] env[65788]: DEBUG nova.objects.instance [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 925.256155] env[65788]: WARNING neutronclient.v2_0.client [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 925.256976] env[65788]: WARNING openstack [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 925.257934] env[65788]: WARNING openstack [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 925.265815] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662670, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.267657] env[65788]: INFO nova.scheduler.client.report [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Deleted allocations for instance 5cef20ed-d9f0-4237-ae9d-da401b953904 [ 925.291453] env[65788]: INFO nova.compute.manager [-] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Took 2.42 seconds to deallocate network for instance. [ 925.343171] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662671, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.423436] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662672, 'name': ReconfigVM_Task, 'duration_secs': 0.193086} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.423436] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.423436] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-223b73b4-88c2-4355-b44c-3a30be8664f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.433058] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 925.433058] env[65788]: value = "task-4662673" [ 925.433058] env[65788]: _type = "Task" [ 925.433058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.443367] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.753733] env[65788]: DEBUG oslo_vmware.api [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662670, 'name': PowerOnVM_Task, 'duration_secs': 0.659779} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.753913] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.754204] env[65788]: INFO nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Took 10.83 seconds to spawn the instance on the hypervisor. [ 925.754405] env[65788]: DEBUG nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 925.755273] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43334224-a79f-459c-b02d-06406c544abe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.778978] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3ccf998c-e5e2-4721-9754-abeb7610c3ce tempest-ServersAdminTestJSON-345343376 tempest-ServersAdminTestJSON-345343376-project-member] Lock "5cef20ed-d9f0-4237-ae9d-da401b953904" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.331s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.843072] env[65788]: DEBUG oslo_vmware.api [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662671, 'name': PowerOnVM_Task, 'duration_secs': 0.569061} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.843790] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.843790] env[65788]: INFO nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Took 13.04 seconds to spawn the instance on the hypervisor. [ 925.843983] env[65788]: DEBUG nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 925.844931] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b457087b-358a-44c4-ba38-cb623f6b10ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.875391] env[65788]: INFO nova.compute.manager [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Took 0.58 seconds to detach 1 volumes for instance. [ 925.946551] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662673, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.193729] env[65788]: WARNING openstack [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.194161] env[65788]: WARNING openstack [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.238161] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1316b2e5-0ce6-4daa-bae1-077cd5dfe8d5 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.240470] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.524s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.240865] env[65788]: DEBUG nova.objects.instance [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lazy-loading 'resources' on Instance uuid 8d7a33e6-08fb-481e-8fac-fdf13b19aa21 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.279668] env[65788]: INFO nova.compute.manager [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Took 43.11 seconds to build instance. [ 926.375639] env[65788]: INFO nova.compute.manager [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Took 43.69 seconds to build instance. [ 926.386885] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.447527] env[65788]: DEBUG oslo_vmware.api [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662673, 'name': PowerOnVM_Task, 'duration_secs': 0.600851} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.449544] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.452476] env[65788]: DEBUG nova.compute.manager [None req-c0bc6cd8-86fb-4011-b699-739dc39d9eb0 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 926.453928] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947370b0-2574-4b3a-be88-8c7bd13f17db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.549461] env[65788]: WARNING neutronclient.v2_0.client [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 926.550539] env[65788]: WARNING openstack [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 926.550967] env[65788]: WARNING openstack [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 926.616617] env[65788]: DEBUG nova.compute.manager [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Received event network-changed-85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 926.616846] env[65788]: DEBUG nova.compute.manager [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Refreshing instance network info cache due to event network-changed-85fbe51b-2508-45d1-a3e3-b6c38fac2060. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 926.616995] env[65788]: DEBUG oslo_concurrency.lockutils [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Acquiring lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.709990] env[65788]: DEBUG nova.network.neutron [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updated VIF entry in instance network info cache for port 85fbe51b-2508-45d1-a3e3-b6c38fac2060. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 926.710442] env[65788]: DEBUG nova.network.neutron [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updating instance_info_cache with network_info: [{"id": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "address": "fa:16:3e:b6:76:23", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85fbe51b-25", "ovs_interfaceid": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 926.782216] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c9df665-2599-4e4b-aefd-ff617e148e8d tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.626s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.889725] env[65788]: DEBUG oslo_concurrency.lockutils [None req-38f99148-ea4a-4913-a541-7e2d152a9c68 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.211s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.124283] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610c67b6-ffb7-484a-b1cc-24ea99bff970 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.133924] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16da39b-5799-480d-bd21-8120084854d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.169667] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fc3802-5809-43ce-8b43-e0aa2cedbac1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.179726] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f77f05-8703-4c08-846d-5efaee9693c7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.197826] env[65788]: DEBUG nova.compute.provider_tree [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.215143] env[65788]: DEBUG oslo_concurrency.lockutils [req-f7c5fdcb-8304-4992-bdc0-678d9d39dbdb req-e1a9245f-91f5-46e3-9697-fc1a7f522022 service nova] Releasing lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 927.215143] env[65788]: DEBUG oslo_concurrency.lockutils [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Acquired lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.215143] env[65788]: DEBUG nova.network.neutron [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Refreshing network info cache for port 85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 927.483172] env[65788]: DEBUG nova.compute.manager [req-3eed5bcf-8f11-47a5-900e-a80c4e94814b req-06f31d4c-f182-4fb6-8739-e8d45d82f956 service nova] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Received event network-vif-deleted-987d7302-963a-4e9f-abaf-6d560099bdbc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 927.702228] env[65788]: DEBUG nova.scheduler.client.report [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 927.718971] env[65788]: WARNING neutronclient.v2_0.client [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 927.718971] env[65788]: WARNING openstack [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 927.719279] env[65788]: WARNING openstack [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 927.772349] env[65788]: INFO nova.compute.manager [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Unrescuing [ 927.772886] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.773109] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.773276] env[65788]: DEBUG nova.network.neutron [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 928.209113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.969s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.212941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.510s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.213253] env[65788]: DEBUG nova.objects.instance [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lazy-loading 'resources' on Instance uuid 7c88e3ef-82d2-46ac-a350-999de8091c01 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.242951] env[65788]: INFO nova.scheduler.client.report [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleted allocations for instance 8d7a33e6-08fb-481e-8fac-fdf13b19aa21 [ 928.280026] env[65788]: WARNING neutronclient.v2_0.client [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 928.280026] env[65788]: WARNING openstack [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 928.280026] env[65788]: WARNING openstack [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 928.754736] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7455299b-7f23-4264-a6bb-18b090f143c2 tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "8d7a33e6-08fb-481e-8fac-fdf13b19aa21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.022s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.102430] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956df65f-813f-46ea-a8c5-6fb06ce97a8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.117128] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e0359c-f44f-4d9e-b9c0-e2f7a4ced501 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.156572] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07aa61f7-f3c7-4398-aee2-4610717b5fa4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.166127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d019de-38a2-4189-9a69-21e1c672ec66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.183363] env[65788]: DEBUG nova.compute.provider_tree [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.446345] env[65788]: WARNING openstack [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.446649] env[65788]: WARNING openstack [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 929.688152] env[65788]: DEBUG nova.scheduler.client.report [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 929.764484] env[65788]: WARNING openstack [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 929.765062] env[65788]: WARNING openstack [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.076413] env[65788]: WARNING neutronclient.v2_0.client [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.077423] env[65788]: WARNING openstack [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.077978] env[65788]: WARNING openstack [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.174541] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.174836] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.194966] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.982s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.197783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.128s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.199653] env[65788]: INFO nova.compute.claims [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.222017] env[65788]: INFO nova.scheduler.client.report [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted allocations for instance 7c88e3ef-82d2-46ac-a350-999de8091c01 [ 930.644122] env[65788]: WARNING neutronclient.v2_0.client [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 930.644873] env[65788]: WARNING openstack [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 930.645264] env[65788]: WARNING openstack [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 930.676871] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 930.734276] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5f273e50-9693-4e3d-992e-f6227f33ad85 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "7c88e3ef-82d2-46ac-a350-999de8091c01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.668s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.818779] env[65788]: DEBUG nova.network.neutron [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updated VIF entry in instance network info cache for port 85fbe51b-2508-45d1-a3e3-b6c38fac2060. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 930.819174] env[65788]: DEBUG nova.network.neutron [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updating instance_info_cache with network_info: [{"id": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "address": "fa:16:3e:b6:76:23", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap85fbe51b-25", "ovs_interfaceid": "85fbe51b-2508-45d1-a3e3-b6c38fac2060", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 931.093442] env[65788]: DEBUG nova.network.neutron [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 931.152890] env[65788]: DEBUG nova.compute.manager [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Received event network-changed-e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 931.153164] env[65788]: DEBUG nova.compute.manager [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Refreshing instance network info cache due to event network-changed-e98f1b15-0a05-4be4-b950-dd7e6749eb8f. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 931.153528] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.153696] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.153888] env[65788]: DEBUG nova.network.neutron [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Refreshing network info cache for port e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 931.216501] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.321893] env[65788]: DEBUG oslo_concurrency.lockutils [req-a87c49c5-0969-429a-8a47-3d156730251b req-51f515ca-a5f8-41a3-aeb0-a666e80a7c60 service nova] Releasing lock "refresh_cache-374564df-70fa-4b89-8f38-e559245e5ebf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.602938] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.603816] env[65788]: DEBUG nova.objects.instance [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'flavor' on Instance uuid 3a06a833-5aaa-4b5d-88b3-8a1d469580af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.619960] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2751083-99d6-435f-9931-0cf6c57527ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.634435] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07a2691-2552-459c-98d9-16325bd2e73b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.671449] env[65788]: WARNING neutronclient.v2_0.client [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 931.672111] env[65788]: WARNING openstack [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 931.675073] env[65788]: WARNING openstack [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 931.681610] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a146c8-538d-4f64-986d-3e24dda9c297 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.690747] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9355a34a-8c40-4d5a-8c6d-b32921ad7319 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.709697] env[65788]: DEBUG nova.compute.provider_tree [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.045128] env[65788]: WARNING openstack [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.045128] env[65788]: WARNING openstack [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.117442] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe88bb84-d11f-45a1-921a-289b014ad88f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.146732] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.147386] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1573e9af-a646-4080-a702-835fe13c01a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.161251] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 932.161251] env[65788]: value = "task-4662674" [ 932.161251] env[65788]: _type = "Task" [ 932.161251] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.179309] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.219391] env[65788]: DEBUG nova.scheduler.client.report [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 932.311434] env[65788]: WARNING neutronclient.v2_0.client [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 932.311733] env[65788]: WARNING openstack [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 932.312108] env[65788]: WARNING openstack [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 932.440630] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.440938] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.441148] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 932.441337] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.441508] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.443868] env[65788]: INFO nova.compute.manager [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Terminating instance [ 932.673601] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662674, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.725555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 932.725952] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 932.728803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.058s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 932.730222] env[65788]: INFO nova.compute.claims [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.795214] env[65788]: DEBUG nova.network.neutron [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updated VIF entry in instance network info cache for port e98f1b15-0a05-4be4-b950-dd7e6749eb8f. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 932.795604] env[65788]: DEBUG nova.network.neutron [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 932.951225] env[65788]: DEBUG nova.compute.manager [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 932.951225] env[65788]: DEBUG nova.compute.manager [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing instance network info cache due to event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 932.951225] env[65788]: DEBUG oslo_concurrency.lockutils [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.951225] env[65788]: DEBUG oslo_concurrency.lockutils [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 932.951225] env[65788]: DEBUG nova.network.neutron [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 932.956345] env[65788]: DEBUG nova.compute.manager [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 932.956345] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.958275] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdd99b4-932e-48ea-956e-234c3f9251f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.968312] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.968610] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6aafdb67-bae0-45c8-8c96-2b9071edde7e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.979997] env[65788]: DEBUG oslo_vmware.api [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 932.979997] env[65788]: value = "task-4662675" [ 932.979997] env[65788]: _type = "Task" [ 932.979997] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.994038] env[65788]: DEBUG oslo_vmware.api [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662675, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.179024] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662674, 'name': PowerOffVM_Task, 'duration_secs': 0.571206} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.179479] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.187144] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfiguring VM instance instance-0000003b to detach disk 2002 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.187585] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2ac9c42-836c-4f43-b906-4a506dbbbaaf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.217833] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 933.217833] env[65788]: value = "task-4662676" [ 933.217833] env[65788]: _type = "Task" [ 933.217833] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.230428] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.235995] env[65788]: DEBUG nova.compute.utils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 933.239971] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 933.240301] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 933.240460] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.240854] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.242077] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.242077] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.298921] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9232f39-4ba3-4ef1-a1cc-c324f434136a req-28a29b34-d5aa-4e2e-9534-ebf4179d56a4 service nova] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 933.461124] env[65788]: WARNING neutronclient.v2_0.client [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 933.461124] env[65788]: WARNING openstack [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 933.461124] env[65788]: WARNING openstack [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 933.495070] env[65788]: DEBUG oslo_vmware.api [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662675, 'name': PowerOffVM_Task, 'duration_secs': 0.199062} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.495521] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.495739] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.496064] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1057c90-0763-47a4-bdc1-01c8dcc6d244 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.604527] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.604990] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.604990] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleting the datastore file [datastore2] 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.606178] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e50e44f0-373e-4b06-8bf8-87052c94a4bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.614854] env[65788]: DEBUG oslo_vmware.api [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 933.614854] env[65788]: value = "task-4662678" [ 933.614854] env[65788]: _type = "Task" [ 933.614854] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.632070] env[65788]: DEBUG oslo_vmware.api [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.649271] env[65788]: DEBUG nova.policy [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2886a02e36704e84ac771401e396629f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd2c665795dc4241a50c46de045f2c80', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 933.655333] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61cf965-1123-4b15-ac77-12a2fb9a78e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.670303] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-037ac4d2-b95e-4a89-9cc1-1cd780316144 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.729252] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30f9519-12c8-4487-a348-e17f220325a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.743478] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662676, 'name': ReconfigVM_Task, 'duration_secs': 0.480974} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.746582] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfigured VM instance instance-0000003b to detach disk 2002 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.748959] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.748959] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51f3502a-2782-4298-be98-f2f933c53eba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.751435] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a11fb3-3e28-4b96-9114-c3e4bea5299b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.758418] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 933.788426] env[65788]: DEBUG nova.compute.provider_tree [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.792026] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 933.792026] env[65788]: value = "task-4662679" [ 933.792026] env[65788]: _type = "Task" [ 933.792026] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.806020] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662679, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.921462] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 933.921688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.127374] env[65788]: DEBUG oslo_vmware.api [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263149} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.134454] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.134874] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.135442] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.135971] env[65788]: INFO nova.compute.manager [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Took 1.18 seconds to destroy the instance on the hypervisor. [ 934.136948] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 934.137718] env[65788]: DEBUG nova.compute.manager [-] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 934.137944] env[65788]: DEBUG nova.network.neutron [-] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 934.138311] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.139440] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.139855] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.164055] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Successfully created port: f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 934.170615] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "37471e59-1809-4df3-8c40-20921d04d18e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.171639] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "37471e59-1809-4df3-8c40-20921d04d18e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.171639] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "37471e59-1809-4df3-8c40-20921d04d18e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.171639] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "37471e59-1809-4df3-8c40-20921d04d18e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.171639] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "37471e59-1809-4df3-8c40-20921d04d18e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.175032] env[65788]: INFO nova.compute.manager [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Terminating instance [ 934.292182] env[65788]: DEBUG nova.scheduler.client.report [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 934.310901] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662679, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.321346] env[65788]: WARNING openstack [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 934.321818] env[65788]: WARNING openstack [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 934.386099] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 934.425033] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 934.682444] env[65788]: DEBUG nova.compute.manager [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 934.682444] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.682444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e50c4a1-c3bc-42da-aac8-6773fe50935d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.688735] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.689352] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec38d86f-80cf-48fe-bacd-7cc5bbabe713 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.699875] env[65788]: DEBUG oslo_vmware.api [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 934.699875] env[65788]: value = "task-4662680" [ 934.699875] env[65788]: _type = "Task" [ 934.699875] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.714833] env[65788]: DEBUG oslo_vmware.api [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662680, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.725646] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Successfully created port: 22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 934.771479] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 934.808285] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 934.808759] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 934.819093] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.593s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 934.822197] env[65788]: INFO nova.compute.claims [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.829464] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 934.831085] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 934.831085] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 934.831085] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 934.831085] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 934.832920] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 934.834077] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.834077] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 934.834077] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 934.834077] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 934.834077] env[65788]: DEBUG nova.virt.hardware [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 934.834817] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0771576e-4c70-4ffb-aeca-8b8d74f5f27c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.844223] env[65788]: DEBUG oslo_vmware.api [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662679, 'name': PowerOnVM_Task, 'duration_secs': 0.632501} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.844223] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.844223] env[65788]: DEBUG nova.compute.manager [None req-fd748f8c-87c7-4a97-917d-3fd000b6ff63 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 934.844459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401ba856-d68b-4fc4-927e-20815195460f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.850490] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cd4f20-0fec-4836-8a5b-40cace0be530 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.959537] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 935.216493] env[65788]: DEBUG oslo_vmware.api [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662680, 'name': PowerOffVM_Task, 'duration_secs': 0.514554} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.217959] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.218272] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.219289] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72570575-3d7f-4c06-8c9e-7525a80c6af5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.287601] env[65788]: WARNING neutronclient.v2_0.client [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.288478] env[65788]: WARNING openstack [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.289034] env[65788]: WARNING openstack [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.300080] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.300566] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.300566] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleting the datastore file [datastore2] 37471e59-1809-4df3-8c40-20921d04d18e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.300818] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d0fc15a1-4e7f-4ffb-a9cb-0c70232a2b29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.309224] env[65788]: DEBUG oslo_vmware.api [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 935.309224] env[65788]: value = "task-4662682" [ 935.309224] env[65788]: _type = "Task" [ 935.309224] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.320189] env[65788]: DEBUG oslo_vmware.api [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662682, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.327028] env[65788]: DEBUG nova.compute.utils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 935.328607] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 935.329023] env[65788]: DEBUG nova.network.neutron [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 935.329562] env[65788]: WARNING neutronclient.v2_0.client [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.329706] env[65788]: WARNING neutronclient.v2_0.client [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.330251] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.330761] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.354762] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.354762] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.464698] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Successfully created port: 0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 935.477987] env[65788]: DEBUG nova.policy [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a5c272c62484b1d8c5a3dc9325631ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94977b75768444138fb914cabddc57c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 935.506150] env[65788]: DEBUG nova.network.neutron [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updated VIF entry in instance network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 935.506150] env[65788]: DEBUG nova.network.neutron [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 935.542039] env[65788]: DEBUG nova.network.neutron [-] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 935.822816] env[65788]: DEBUG oslo_vmware.api [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662682, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152477} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.824890] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.824976] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.826041] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.826041] env[65788]: INFO nova.compute.manager [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 935.826041] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 935.826041] env[65788]: DEBUG nova.compute.manager [-] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 935.826041] env[65788]: DEBUG nova.network.neutron [-] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 935.826259] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 935.826900] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 935.826998] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 935.840138] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 935.862706] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.862918] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.863404] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.865492] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.865492] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.865492] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.865492] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 935.865492] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 935.897104] env[65788]: DEBUG nova.network.neutron [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Successfully created port: eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 936.006532] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.006769] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.008700] env[65788]: DEBUG oslo_concurrency.lockutils [req-5f6eed21-38fd-4987-b930-12678eca4dd9 req-e6b23020-2569-45ca-8b10-26bd8d153c4e service nova] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 936.043414] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 936.045846] env[65788]: INFO nova.compute.manager [-] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Took 1.91 seconds to deallocate network for instance. [ 936.102415] env[65788]: DEBUG nova.compute.manager [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 936.239640] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2595533-f06c-4bc0-824c-282def04629b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.249219] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b899cb-e01f-49ca-b75a-8dafe6c02fa7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.286520] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a7d6a4-1d60-4610-a797-a8b9106e9698 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.295058] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61833cdc-9be0-4ea8-86b4-b76657437651 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.311169] env[65788]: DEBUG nova.compute.provider_tree [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.366609] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.510383] env[65788]: DEBUG nova.compute.utils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 936.552913] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.621846] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.816201] env[65788]: DEBUG nova.scheduler.client.report [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 936.858493] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 936.884932] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 936.885283] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 936.885489] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 936.885708] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 936.885854] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 936.885998] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 936.886222] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.886379] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 936.886556] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 936.886718] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 936.886886] env[65788]: DEBUG nova.virt.hardware [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 936.887781] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4323373b-aa3e-4ac5-83c0-fa35b8a2d7e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.896739] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e04e6ca-ef87-46b6-894c-5eca1dca3f91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.013806] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.119908] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Successfully updated port: f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 937.322167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.322779] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 937.326427] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.940s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.326769] env[65788]: DEBUG nova.objects.instance [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'resources' on Instance uuid 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.624498] env[65788]: DEBUG nova.network.neutron [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Successfully updated port: eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 937.831401] env[65788]: DEBUG nova.compute.utils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 937.836816] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 937.836816] env[65788]: DEBUG nova.network.neutron [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 937.836816] env[65788]: WARNING neutronclient.v2_0.client [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.837782] env[65788]: WARNING neutronclient.v2_0.client [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 937.837782] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 937.837884] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 937.919479] env[65788]: DEBUG nova.network.neutron [-] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 938.111794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.114461] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.114461] env[65788]: INFO nova.compute.manager [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Attaching volume 1294ba3f-809b-4d9c-be6d-9441e471e9ea to /dev/sdb [ 938.129167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.129382] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 938.129567] env[65788]: DEBUG nova.network.neutron [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 938.154507] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8b4e82-7284-4f57-a66f-b8e3c79c900f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.164749] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02285f9d-eaf3-42e6-8f8b-c3f7db5f10d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.185770] env[65788]: DEBUG nova.virt.block_device [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating existing volume attachment record: 62eee814-dae8-4b87-a327-cd915fb4cfdb {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 938.221620] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a851f34e-1b9d-403c-a7ab-dfe1ebfc253c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.233360] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8c4000-ede3-4a96-b71a-f2f5dcf9fd9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.271620] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8f9c49-421c-4821-9ac5-92d80be0b8c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.277612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.277858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 938.284695] env[65788]: DEBUG nova.policy [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e095550d08426cb59f11d2295b9a60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '246c0426008f478aa245d006c3c129eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 938.292074] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb100c91-5432-421f-8fa7-73e7f60275ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.308333] env[65788]: DEBUG nova.compute.provider_tree [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.346012] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 938.421333] env[65788]: INFO nova.compute.manager [-] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Took 2.60 seconds to deallocate network for instance. [ 938.634944] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 938.636185] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 938.781171] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 938.811604] env[65788]: DEBUG nova.scheduler.client.report [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 938.928232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.143191] env[65788]: DEBUG nova.network.neutron [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Successfully created port: 00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 939.231417] env[65788]: DEBUG nova.network.neutron [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 939.254560] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Successfully updated port: 22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 939.273250] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.273250] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.305765] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.318354] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.323171] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.105s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.323990] env[65788]: INFO nova.compute.claims [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.358336] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 939.362463] env[65788]: INFO nova.scheduler.client.report [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocations for instance 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca [ 939.395131] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 939.395394] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 939.395548] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 939.395831] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 939.396661] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 939.396661] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 939.396661] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.396661] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 939.396661] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 939.396984] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 939.396984] env[65788]: DEBUG nova.virt.hardware [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 939.398134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95211af9-2ca1-4f5e-a062-c4fd638cba1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.407955] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61390551-63aa-47c5-92e1-6bd1559d0a62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.486798] env[65788]: WARNING neutronclient.v2_0.client [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 939.487510] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 939.487855] env[65788]: WARNING openstack [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 939.704055] env[65788]: DEBUG nova.compute.manager [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 939.705518] env[65788]: DEBUG nova.compute.manager [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing instance network info cache due to event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 939.705518] env[65788]: DEBUG oslo_concurrency.lockutils [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.705518] env[65788]: DEBUG oslo_concurrency.lockutils [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.705518] env[65788]: DEBUG nova.network.neutron [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 939.713740] env[65788]: DEBUG nova.network.neutron [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 939.875492] env[65788]: DEBUG oslo_concurrency.lockutils [None req-36fbed3c-de44-4f44-8f81-444266a9452d tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.848s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.208095] env[65788]: WARNING neutronclient.v2_0.client [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 940.208863] env[65788]: WARNING openstack [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 940.209230] env[65788]: WARNING openstack [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 940.216945] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.217300] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance network_info: |[{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 940.217760] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:23:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a0a76279-3c11-4bef-b124-2a2ee13fa377', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eba2c4b7-13de-4fe3-95fb-86d9494cd97a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.225490] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating folder: Project (94977b75768444138fb914cabddc57c5). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 940.226253] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1d800e6-5b9e-45a4-9114-e896b604f192 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.237893] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created folder: Project (94977b75768444138fb914cabddc57c5) in parent group-v910111. [ 940.238107] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating folder: Instances. Parent ref: group-v910333. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 940.238384] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2632ae10-372b-45fd-80d4-17c098fffd19 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.252146] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created folder: Instances in parent group-v910333. [ 940.252516] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 940.252758] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.252996] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81069e53-bd06-4b96-949e-83b4c945d2b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.274358] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.274358] env[65788]: value = "task-4662688" [ 940.274358] env[65788]: _type = "Task" [ 940.274358] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.284187] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662688, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.474716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.474716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.668423] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2633d9-7bbf-4aca-b332-7647685692dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.678722] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2405fdb3-c9c7-4ec7-a677-cbf42c21b8fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.715169] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c1f00f-f975-44a3-a3e2-1112513df0e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.724877] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fb5496-788a-437b-b85d-9ea786f4ddad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.742448] env[65788]: DEBUG nova.compute.provider_tree [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.786636] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662688, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.824906] env[65788]: DEBUG nova.network.neutron [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Successfully updated port: 00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 940.976226] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 941.246446] env[65788]: DEBUG nova.scheduler.client.report [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 941.288433] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662688, 'name': CreateVM_Task, 'duration_secs': 0.525368} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.288642] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.289382] env[65788]: WARNING neutronclient.v2_0.client [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.289807] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.289977] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.290321] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 941.290736] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abb2aa11-0b44-4337-9b6d-20869cd76f11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.297557] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 941.297557] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524e3833-a5cf-eb63-07a6-a7ba653f348f" [ 941.297557] env[65788]: _type = "Task" [ 941.297557] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.311406] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524e3833-a5cf-eb63-07a6-a7ba653f348f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.330549] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "refresh_cache-e0f7c849-315c-4247-a840-d388d48746b9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.330549] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "refresh_cache-e0f7c849-315c-4247-a840-d388d48746b9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.330549] env[65788]: DEBUG nova.network.neutron [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 941.337208] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Successfully updated port: 0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 941.442421] env[65788]: WARNING openstack [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.442800] env[65788]: WARNING openstack [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.452992] env[65788]: DEBUG nova.compute.manager [req-a340e6dc-fdff-460b-a66b-64f1a2e5efe1 req-393ef46f-3049-465c-99a2-e52631461318 service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-vif-plugged-f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 941.453228] env[65788]: DEBUG oslo_concurrency.lockutils [req-a340e6dc-fdff-460b-a66b-64f1a2e5efe1 req-393ef46f-3049-465c-99a2-e52631461318 service nova] Acquiring lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.454325] env[65788]: DEBUG oslo_concurrency.lockutils [req-a340e6dc-fdff-460b-a66b-64f1a2e5efe1 req-393ef46f-3049-465c-99a2-e52631461318 service nova] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.455407] env[65788]: DEBUG oslo_concurrency.lockutils [req-a340e6dc-fdff-460b-a66b-64f1a2e5efe1 req-393ef46f-3049-465c-99a2-e52631461318 service nova] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.455407] env[65788]: DEBUG nova.compute.manager [req-a340e6dc-fdff-460b-a66b-64f1a2e5efe1 req-393ef46f-3049-465c-99a2-e52631461318 service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] No waiting events found dispatching network-vif-plugged-f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 941.455407] env[65788]: WARNING nova.compute.manager [req-a340e6dc-fdff-460b-a66b-64f1a2e5efe1 req-393ef46f-3049-465c-99a2-e52631461318 service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received unexpected event network-vif-plugged-f795ad8a-6471-483d-a6b9-a381bad1c376 for instance with vm_state building and task_state spawning. [ 941.506868] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.542721] env[65788]: WARNING neutronclient.v2_0.client [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 941.543433] env[65788]: WARNING openstack [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.544232] env[65788]: WARNING openstack [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.753588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.754156] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 941.757485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.798s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.758930] env[65788]: INFO nova.compute.claims [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.813886] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524e3833-a5cf-eb63-07a6-a7ba653f348f, 'name': SearchDatastore_Task, 'duration_secs': 0.011918} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.814200] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.814345] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.814770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.814770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.814882] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.815210] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13393d77-dd2e-4aaf-a09e-afcada35ebe3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.827792] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.827792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.828144] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-709de869-5ac0-4468-86e7-edda49769ea4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.835813] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 941.836199] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 941.849923] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 941.849923] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525301a0-6825-95a4-a05f-9224f7ba8799" [ 941.849923] env[65788]: _type = "Task" [ 941.849923] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.849923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.850202] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.850261] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 941.864785] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525301a0-6825-95a4-a05f-9224f7ba8799, 'name': SearchDatastore_Task, 'duration_secs': 0.01089} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.866241] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac9b02af-d1b8-4d9c-9d35-f6ede51284b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.873992] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 941.873992] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ba7999-fcc0-64dd-fbd7-b106eb7d5b13" [ 941.873992] env[65788]: _type = "Task" [ 941.873992] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.884128] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ba7999-fcc0-64dd-fbd7-b106eb7d5b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.974956] env[65788]: DEBUG nova.network.neutron [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updated VIF entry in instance network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 941.975385] env[65788]: DEBUG nova.network.neutron [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 942.022657] env[65788]: DEBUG nova.network.neutron [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 942.209857] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.210289] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.263785] env[65788]: DEBUG nova.compute.utils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 942.272361] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 942.273721] env[65788]: DEBUG nova.network.neutron [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 942.273721] env[65788]: WARNING neutronclient.v2_0.client [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.273721] env[65788]: WARNING neutronclient.v2_0.client [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.274598] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.274598] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.340671] env[65788]: DEBUG nova.policy [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca13693271604f5790bdc8e19aa208a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7aa1dec53aa4436daa0f3223b92f9b79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 942.359390] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.359390] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.386328] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ba7999-fcc0-64dd-fbd7-b106eb7d5b13, 'name': SearchDatastore_Task, 'duration_secs': 0.010288} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.386770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.387052] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.387330] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d025b3cf-f4bc-42c1-bbab-a5437276707f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.392802] env[65788]: WARNING neutronclient.v2_0.client [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.393543] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.393989] env[65788]: WARNING openstack [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.408902] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 942.408902] env[65788]: value = "task-4662690" [ 942.408902] env[65788]: _type = "Task" [ 942.408902] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.424582] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662690, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.463909] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 942.480210] env[65788]: DEBUG oslo_concurrency.lockutils [req-1e0f816d-cf77-40d2-b517-4209c6e91228 req-9c6dea68-7dd5-46a2-b719-b25167e587ec service nova] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.608578] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.609071] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.642546] env[65788]: DEBUG nova.network.neutron [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Updating instance_info_cache with network_info: [{"id": "00af96db-551c-48f3-9b82-c272038d58a8", "address": "fa:16:3e:21:93:70", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00af96db-55", "ovs_interfaceid": "00af96db-551c-48f3-9b82-c272038d58a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 942.754542] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 942.756109] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910332', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'name': 'volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd8103991-fd1f-4a57-81ce-1a47dc4defe0', 'attached_at': '', 'detached_at': '', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'serial': '1294ba3f-809b-4d9c-be6d-9441e471e9ea'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 942.756377] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d6d6e9-5b9b-4b47-a28d-42658a101a55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.783950] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 942.793298] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ea2709-4def-4da3-8a6e-fee3fefcbd44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.825973] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea/volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.833469] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 942.833469] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 942.833469] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 942.842131] env[65788]: DEBUG nova.network.neutron [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Successfully created port: b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 942.844593] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e92ffa88-7da2-4d7d-b56d-ec648404c616 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.869837] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 942.869837] env[65788]: value = "task-4662691" [ 942.869837] env[65788]: _type = "Task" [ 942.869837] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.882580] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.921819] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662690, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.004030] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Received event network-vif-deleted-02faaa1a-1e5f-46a3-ba39-2e846e4ce4aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 943.004030] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-vif-plugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 943.004030] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.004407] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 943.004407] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 943.004538] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] No waiting events found dispatching network-vif-plugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 943.004713] env[65788]: WARNING nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received unexpected event network-vif-plugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a for instance with vm_state building and task_state spawning. [ 943.004882] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 943.005036] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing instance network info cache due to event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 943.006127] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.006127] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.006127] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 943.093530] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.095445] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.152053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "refresh_cache-e0f7c849-315c-4247-a840-d388d48746b9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 943.152053] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Instance network_info: |[{"id": "00af96db-551c-48f3-9b82-c272038d58a8", "address": "fa:16:3e:21:93:70", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00af96db-55", "ovs_interfaceid": "00af96db-551c-48f3-9b82-c272038d58a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 943.152318] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:93:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00af96db-551c-48f3-9b82-c272038d58a8', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.160730] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 943.160871] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 943.161080] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51ff2ee6-4f1d-43d7-8988-03fb0af08729 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.186816] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.186816] env[65788]: value = "task-4662692" [ 943.186816] env[65788]: _type = "Task" [ 943.186816] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.204398] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662692, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.289873] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.290764] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.291256] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.341718] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d711c263-c5a6-4793-8ba8-dd16626bb066 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.350720] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2916d4-9770-4cc6-abf2-f3955e48bfa7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.393069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8aa991-302d-4757-8dcf-8ca6ba263756 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.406690] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16afe6da-630e-4ffe-b987-f2a5cd835e34 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.415010] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662691, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.426278] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662690, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.437266] env[65788]: DEBUG nova.compute.provider_tree [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.513707] env[65788]: WARNING neutronclient.v2_0.client [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.514647] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.515184] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.656372] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.656781] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.698936] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662692, 'name': CreateVM_Task, 'duration_secs': 0.48484} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.698936] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.699391] env[65788]: WARNING neutronclient.v2_0.client [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.699758] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.699906] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 943.701040] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 943.701040] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9d835ef-6d2a-488e-8527-35419595bf4c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.706147] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 943.706147] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fdcda2-8d78-126e-e73c-4d96d362c270" [ 943.706147] env[65788]: _type = "Task" [ 943.706147] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.715599] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fdcda2-8d78-126e-e73c-4d96d362c270, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.783334] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.783763] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.814922] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 943.815640] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 943.816019] env[65788]: WARNING openstack [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 943.831431] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 943.860450] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 943.860709] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 943.860863] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 943.861057] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 943.861272] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 943.861385] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 943.861617] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.861790] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 943.861958] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 943.862137] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 943.862312] env[65788]: DEBUG nova.virt.hardware [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 943.863234] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87333c3-ccd5-4ae5-9fda-2b0537c40157 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.877430] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54e8b20-f38e-4f86-9244-b5740a2c9ccc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.902938] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662691, 'name': ReconfigVM_Task, 'duration_secs': 0.593552} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.903096] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Reconfigured VM instance instance-00000042 to attach disk [datastore1] volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea/volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.907934] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6561fe0-131f-4a64-ba88-4d927cf621bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.929102] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662690, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.930663] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 943.930663] env[65788]: value = "task-4662693" [ 943.930663] env[65788]: _type = "Task" [ 943.930663] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.940797] env[65788]: DEBUG nova.scheduler.client.report [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 943.951539] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662693, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.063280] env[65788]: WARNING neutronclient.v2_0.client [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.064798] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.064922] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.115291] env[65788]: DEBUG nova.network.neutron [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updating instance_info_cache with network_info: [{"id": "f795ad8a-6471-483d-a6b9-a381bad1c376", "address": "fa:16:3e:38:9b:0a", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf795ad8a-64", "ovs_interfaceid": "f795ad8a-6471-483d-a6b9-a381bad1c376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "address": "fa:16:3e:7f:6a:c0", "network": {"id": "69f9d02f-97b1-4b21-80da-80e584d26f8b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1353042792", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.32", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22f6c5a6-c6", "ovs_interfaceid": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "address": "fa:16:3e:8e:31:38", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9e89f7-41", "ovs_interfaceid": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 944.220223] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fdcda2-8d78-126e-e73c-4d96d362c270, 'name': SearchDatastore_Task, 'duration_secs': 0.010409} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.220223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.220223] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 944.220223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.220223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.220616] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 944.220952] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42928253-5b38-455d-867d-a104aee93b77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.231361] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 944.231561] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 944.232399] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9492959-c788-42ea-b873-c1e8625ce905 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.239550] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 944.239550] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223672a-18c5-3bf5-c10b-f15d12666f43" [ 944.239550] env[65788]: _type = "Task" [ 944.239550] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.248314] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223672a-18c5-3bf5-c10b-f15d12666f43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.272231] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updated VIF entry in instance network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 944.272672] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 944.430778] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662690, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.647582} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.431888] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.431888] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.434993] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d009324f-a1d9-4e89-8863-5e482841c149 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.445351] env[65788]: DEBUG oslo_vmware.api [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662693, 'name': ReconfigVM_Task, 'duration_secs': 0.154239} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.445351] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910332', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'name': 'volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd8103991-fd1f-4a57-81ce-1a47dc4defe0', 'attached_at': '', 'detached_at': '', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'serial': '1294ba3f-809b-4d9c-be6d-9441e471e9ea'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 944.446212] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 944.446212] env[65788]: value = "task-4662694" [ 944.446212] env[65788]: _type = "Task" [ 944.446212] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.446869] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.689s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.448674] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 944.451779] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.085s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.451967] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.452136] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 944.452422] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.900s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.452680] env[65788]: DEBUG nova.objects.instance [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lazy-loading 'resources' on Instance uuid 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.464180] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f767fa31-4af2-40c4-8863-7371ac6ee543 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.467264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.467484] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.475787] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662694, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.479226] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015461c2-f2a8-45fb-bfea-cb954e7c9481 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.495472] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53599a79-7601-45d9-8e42-9af22fb65167 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.505333] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3f938e-8f65-4579-8d48-cc65c821acd9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.538702] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178304MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 944.538853] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.559147] env[65788]: DEBUG nova.network.neutron [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Successfully updated port: b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 944.617981] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Releasing lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.618527] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Instance network_info: |[{"id": "f795ad8a-6471-483d-a6b9-a381bad1c376", "address": "fa:16:3e:38:9b:0a", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf795ad8a-64", "ovs_interfaceid": "f795ad8a-6471-483d-a6b9-a381bad1c376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "address": "fa:16:3e:7f:6a:c0", "network": {"id": "69f9d02f-97b1-4b21-80da-80e584d26f8b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1353042792", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.32", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22f6c5a6-c6", "ovs_interfaceid": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "address": "fa:16:3e:8e:31:38", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9e89f7-41", "ovs_interfaceid": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 944.619069] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:9b:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e27fd35-1d7b-4358-92d5-4d34da27b992', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f795ad8a-6471-483d-a6b9-a381bad1c376', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:6a:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22f6c5a6-c617-44cd-9f61-61773c5a18f6', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:31:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e27fd35-1d7b-4358-92d5-4d34da27b992', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c9e89f7-4194-4cd0-86aa-72b4f8655f1e', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.630230] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Creating folder: Project (dd2c665795dc4241a50c46de045f2c80). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.630598] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bb25357-d3bb-47d3-984d-a786b1926e54 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.642978] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Created folder: Project (dd2c665795dc4241a50c46de045f2c80) in parent group-v910111. [ 944.642978] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Creating folder: Instances. Parent ref: group-v910337. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.642978] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06dd3fbd-7197-4b28-8b9e-e45abe6c8cce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.651847] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Created folder: Instances in parent group-v910337. [ 944.652107] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 944.652324] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 944.652552] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bea048c2-27c6-48cc-b440-834baad88086 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.676911] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.676911] env[65788]: value = "task-4662697" [ 944.676911] env[65788]: _type = "Task" [ 944.676911] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.685696] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662697, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.690395] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-changed-f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 944.690904] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Refreshing instance network info cache due to event network-changed-f795ad8a-6471-483d-a6b9-a381bad1c376. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 944.690904] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquiring lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.691081] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquired lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 944.691119] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Refreshing network info cache for port f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 944.751039] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223672a-18c5-3bf5-c10b-f15d12666f43, 'name': SearchDatastore_Task, 'duration_secs': 0.010131} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.751884] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-679d51bf-2002-4d09-aacf-1adf1aaa9567 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.759539] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 944.759539] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52171086-9e1e-9252-088b-c85a83a29db5" [ 944.759539] env[65788]: _type = "Task" [ 944.759539] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.769051] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52171086-9e1e-9252-088b-c85a83a29db5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.775731] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.776034] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-vif-plugged-22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 944.776280] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquiring lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 944.776491] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.776674] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.776846] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] No waiting events found dispatching network-vif-plugged-22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 944.777060] env[65788]: WARNING nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received unexpected event network-vif-plugged-22f6c5a6-c617-44cd-9f61-61773c5a18f6 for instance with vm_state building and task_state spawning. [ 944.777187] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-changed-22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 944.777341] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Refreshing instance network info cache due to event network-changed-22f6c5a6-c617-44cd-9f61-61773c5a18f6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 944.777504] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquiring lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.962034] env[65788]: DEBUG nova.compute.utils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 944.963022] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 944.963022] env[65788]: DEBUG nova.network.neutron [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 944.963241] env[65788]: WARNING neutronclient.v2_0.client [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.963527] env[65788]: WARNING neutronclient.v2_0.client [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 944.968344] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 944.968918] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 944.984059] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 944.990632] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662694, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064022} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.991164] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.992021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-898ce5c2-7de4-4ad5-8e74-1a26b7cc92b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.021845] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.028395] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b22f3869-541b-48c0-9ef5-cb0c33c6af1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.051453] env[65788]: DEBUG nova.policy [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0dbfe01341d44724b9f1eee7ea9d5e9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99b3e41827194188800da4e0544bc19f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 945.056382] env[65788]: DEBUG nova.objects.instance [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.064556] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.064556] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.064556] env[65788]: DEBUG nova.network.neutron [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 945.067749] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 945.067749] env[65788]: value = "task-4662698" [ 945.067749] env[65788]: _type = "Task" [ 945.067749] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.084668] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662698, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.188450] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662697, 'name': CreateVM_Task, 'duration_secs': 0.485981} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.188450] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.191907] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.192349] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.192871] env[65788]: WARNING neutronclient.v2_0.client [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.193235] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.193397] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.193757] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 945.194781] env[65788]: WARNING neutronclient.v2_0.client [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.195158] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.195510] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.202742] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31e05bc0-ac56-4ec5-8c4d-c2cb729eca31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.212842] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 945.212842] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b25c66-6cb9-52ad-aab1-0ddb7909bd55" [ 945.212842] env[65788]: _type = "Task" [ 945.212842] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.221464] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b25c66-6cb9-52ad-aab1-0ddb7909bd55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.277379] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52171086-9e1e-9252-088b-c85a83a29db5, 'name': SearchDatastore_Task, 'duration_secs': 0.010026} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.277629] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.277773] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] e0f7c849-315c-4247-a840-d388d48746b9/e0f7c849-315c-4247-a840-d388d48746b9.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 945.278056] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-343bc023-0434-45d3-b221-3225711c89f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.289423] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 945.289423] env[65788]: value = "task-4662699" [ 945.289423] env[65788]: _type = "Task" [ 945.289423] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.301288] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.400886] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.400886] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.465363] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 945.469636] env[65788]: DEBUG nova.network.neutron [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Successfully created port: 6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 945.506068] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78883d9b-de11-46f3-9bea-7b0383b3531f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.514868] env[65788]: WARNING neutronclient.v2_0.client [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 945.514868] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.514868] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.522652] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.525675] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dabc843a-e42b-4083-9da4-5142c7609d58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.569252] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.569828] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.578971] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df24a45-cab4-4b93-b275-e4480139f2a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.587146] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ddeba87f-eb9c-41ea-a7c0-e88d9fada6b2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.474s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.599133] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662698, 'name': ReconfigVM_Task, 'duration_secs': 0.310249} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.607825] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.607825] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a21d3c8-c705-4ddb-9556-e449e3a0297e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.607825] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a1467b-f765-4754-bb2b-0407bad8ccb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.626734] env[65788]: DEBUG nova.compute.provider_tree [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.633932] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 945.633932] env[65788]: value = "task-4662700" [ 945.633932] env[65788]: _type = "Task" [ 945.633932] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.648917] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662700, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.725369] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b25c66-6cb9-52ad-aab1-0ddb7909bd55, 'name': SearchDatastore_Task, 'duration_secs': 0.009234} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.725672] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 945.725878] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.726198] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.726365] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.726630] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.727129] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a1ebdeb-e3a4-4a05-890f-c0a56654c55f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.742165] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.742613] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.743521] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adf4f09b-add0-4210-a8f0-03ad7749208e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.751079] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 945.751079] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e22b89-a906-331b-559f-2c5342fbc509" [ 945.751079] env[65788]: _type = "Task" [ 945.751079] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.763802] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e22b89-a906-331b-559f-2c5342fbc509, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.779585] env[65788]: DEBUG nova.network.neutron [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 945.801096] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662699, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.815231] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updated VIF entry in instance network info cache for port f795ad8a-6471-483d-a6b9-a381bad1c376. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 945.815698] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updating instance_info_cache with network_info: [{"id": "f795ad8a-6471-483d-a6b9-a381bad1c376", "address": "fa:16:3e:38:9b:0a", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf795ad8a-64", "ovs_interfaceid": "f795ad8a-6471-483d-a6b9-a381bad1c376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "address": "fa:16:3e:7f:6a:c0", "network": {"id": "69f9d02f-97b1-4b21-80da-80e584d26f8b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1353042792", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.32", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22f6c5a6-c6", "ovs_interfaceid": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "address": "fa:16:3e:8e:31:38", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9e89f7-41", "ovs_interfaceid": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 945.908439] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 945.908439] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 945.927321] env[65788]: DEBUG nova.compute.manager [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 945.927321] env[65788]: DEBUG nova.compute.manager [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing instance network info cache due to event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 945.927321] env[65788]: DEBUG oslo_concurrency.lockutils [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.927321] env[65788]: DEBUG oslo_concurrency.lockutils [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 945.927321] env[65788]: DEBUG nova.network.neutron [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 946.087685] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.087812] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.131586] env[65788]: WARNING neutronclient.v2_0.client [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.133079] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.133079] env[65788]: WARNING openstack [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.144571] env[65788]: DEBUG nova.scheduler.client.report [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.159969] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662700, 'name': Rename_Task, 'duration_secs': 0.284762} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.160318] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.160581] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-019c108d-f218-4fc1-8a00-f40d33e7e094 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.168778] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 946.168778] env[65788]: value = "task-4662701" [ 946.168778] env[65788]: _type = "Task" [ 946.168778] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.177436] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662701, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.264260] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e22b89-a906-331b-559f-2c5342fbc509, 'name': SearchDatastore_Task, 'duration_secs': 0.04224} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.265165] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd051980-0cc4-4276-96e2-f48319eef308 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.271455] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 946.271455] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527e8dd2-3d75-5158-4fea-07f7236d316c" [ 946.271455] env[65788]: _type = "Task" [ 946.271455] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.281894] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527e8dd2-3d75-5158-4fea-07f7236d316c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.301538] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520414} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.302223] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] e0f7c849-315c-4247-a840-d388d48746b9/e0f7c849-315c-4247-a840-d388d48746b9.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 946.302223] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 946.302434] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2608a100-edcc-4cdc-9ff4-20a260a15536 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.310973] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 946.310973] env[65788]: value = "task-4662702" [ 946.310973] env[65788]: _type = "Task" [ 946.310973] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.322601] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Releasing lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.322601] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Received event network-vif-deleted-ef83244d-23de-4589-bab0-ea354e648ac2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 946.322855] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Received event network-vif-plugged-00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 946.323768] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquiring lock "e0f7c849-315c-4247-a840-d388d48746b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 946.323768] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Lock "e0f7c849-315c-4247-a840-d388d48746b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.323768] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Lock "e0f7c849-315c-4247-a840-d388d48746b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.323768] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] No waiting events found dispatching network-vif-plugged-00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 946.323768] env[65788]: WARNING nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Received unexpected event network-vif-plugged-00af96db-551c-48f3-9b82-c272038d58a8 for instance with vm_state building and task_state spawning. [ 946.324075] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Received event network-changed-00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 946.324126] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Refreshing instance network info cache due to event network-changed-00af96db-551c-48f3-9b82-c272038d58a8. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 946.324297] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquiring lock "refresh_cache-e0f7c849-315c-4247-a840-d388d48746b9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.324428] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquired lock "refresh_cache-e0f7c849-315c-4247-a840-d388d48746b9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.324588] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Refreshing network info cache for port 00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 946.325976] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.326248] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquired lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 946.326427] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Refreshing network info cache for port 22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 946.333770] env[65788]: DEBUG nova.network.neutron [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updating instance_info_cache with network_info: [{"id": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "address": "fa:16:3e:37:b2:1d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62b03bf-7b", "ovs_interfaceid": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 946.429153] env[65788]: WARNING neutronclient.v2_0.client [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.429968] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.430384] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.485038] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 946.522051] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 946.522331] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 946.522484] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 946.522699] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 946.522848] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 946.522991] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 946.523215] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.523367] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 946.523535] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 946.523739] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 946.523953] env[65788]: DEBUG nova.virt.hardware [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 946.524873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09611a02-1145-4ad4-9277-1551ffcc6995 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.539031] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a21185c-89a9-4276-895a-b7cb7a8e4efb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.590971] env[65788]: DEBUG nova.compute.manager [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 946.650453] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.198s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 946.653212] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 10.031s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 946.680143] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662701, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.683693] env[65788]: INFO nova.scheduler.client.report [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleted allocations for instance 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a [ 946.700567] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.702047] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.783854] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527e8dd2-3d75-5158-4fea-07f7236d316c, 'name': SearchDatastore_Task, 'duration_secs': 0.014165} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.783854] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.783854] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8/6a69d645-f89d-43b6-8bdc-b12ce0d88ca8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.783854] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-703ca80c-3a0a-41e5-b633-739704918137 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.791692] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 946.791692] env[65788]: value = "task-4662703" [ 946.791692] env[65788]: _type = "Task" [ 946.791692] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.804537] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.822045] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103827} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.822267] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 946.823111] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e09bdb7-1ee7-4a07-8654-94fa5ade2381 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.828659] env[65788]: WARNING neutronclient.v2_0.client [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.829335] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.829671] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.849402] env[65788]: WARNING neutronclient.v2_0.client [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 946.850975] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 946.854023] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 946.859249] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 946.859784] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Instance network_info: |[{"id": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "address": "fa:16:3e:37:b2:1d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62b03bf-7b", "ovs_interfaceid": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 946.868807] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] e0f7c849-315c-4247-a840-d388d48746b9/e0f7c849-315c-4247-a840-d388d48746b9.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.869638] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:b2:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b62b03bf-7b53-4c01-880d-43cd9bf586f6', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.877367] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Creating folder: Project (7aa1dec53aa4436daa0f3223b92f9b79). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.877750] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5dc8e037-05d6-4070-970f-cd896eaf2189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.892810] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39ab6818-58c3-4977-89e4-2933416f4223 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.901365] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 946.901365] env[65788]: value = "task-4662705" [ 946.901365] env[65788]: _type = "Task" [ 946.901365] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.906446] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Created folder: Project (7aa1dec53aa4436daa0f3223b92f9b79) in parent group-v910111. [ 946.906632] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Creating folder: Instances. Parent ref: group-v910340. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.907290] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bf94a9c-dca2-4b5e-96d5-8546543fcf71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.913858] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.919870] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Created folder: Instances in parent group-v910340. [ 946.919870] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 946.919870] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.919870] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41059077-064e-4b5c-978e-e77543ac4263 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.946318] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.946318] env[65788]: value = "task-4662707" [ 946.946318] env[65788]: _type = "Task" [ 946.946318] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.957677] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662707, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.107709] env[65788]: WARNING neutronclient.v2_0.client [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.108477] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.109199] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.118575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.158585] env[65788]: INFO nova.compute.claims [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.170930] env[65788]: DEBUG nova.network.neutron [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Successfully updated port: 6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 947.190471] env[65788]: DEBUG oslo_vmware.api [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662701, 'name': PowerOnVM_Task, 'duration_secs': 0.910846} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.195282] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.196406] env[65788]: INFO nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Took 10.34 seconds to spawn the instance on the hypervisor. [ 947.196406] env[65788]: DEBUG nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 947.202306] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dfee7c07-936c-4b27-a6a6-d62581aee85b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.761s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.204181] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59464110-4b9c-43a6-98be-a9530df34bc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.309485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.309762] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.309969] env[65788]: DEBUG nova.compute.manager [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 947.312195] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1d7dfe-779f-43f8-913f-e19e42d307a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.322912] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662703, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.324591] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.325132] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.335680] env[65788]: DEBUG nova.compute.manager [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 947.336358] env[65788]: DEBUG nova.objects.instance [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.341291] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.341291] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.352454] env[65788]: DEBUG nova.network.neutron [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updated VIF entry in instance network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 947.352847] env[65788]: DEBUG nova.network.neutron [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.414477] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662705, 'name': ReconfigVM_Task, 'duration_secs': 0.409003} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.416098] env[65788]: WARNING neutronclient.v2_0.client [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.416825] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.417283] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.425931] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Reconfigured VM instance instance-0000004c to attach disk [datastore1] e0f7c849-315c-4247-a840-d388d48746b9/e0f7c849-315c-4247-a840-d388d48746b9.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 947.426944] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f23e631-16ab-4aa3-a9fd-7d1ab5547cb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.435962] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 947.435962] env[65788]: value = "task-4662708" [ 947.435962] env[65788]: _type = "Task" [ 947.435962] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.443234] env[65788]: WARNING neutronclient.v2_0.client [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.443982] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 947.444364] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 947.460977] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662708, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.474529] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662707, 'name': CreateVM_Task, 'duration_secs': 0.398784} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.474727] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.475576] env[65788]: WARNING neutronclient.v2_0.client [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 947.475794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.475977] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.476641] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 947.476716] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f312ef-5af3-485f-a815-53cc79ccd52e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.491342] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 947.491342] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523e1eed-5044-5987-ca58-01f16f9d7ecd" [ 947.491342] env[65788]: _type = "Task" [ 947.491342] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.502186] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523e1eed-5044-5987-ca58-01f16f9d7ecd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.534100] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Updated VIF entry in instance network info cache for port 00af96db-551c-48f3-9b82-c272038d58a8. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 947.535895] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Updating instance_info_cache with network_info: [{"id": "00af96db-551c-48f3-9b82-c272038d58a8", "address": "fa:16:3e:21:93:70", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00af96db-55", "ovs_interfaceid": "00af96db-551c-48f3-9b82-c272038d58a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.548614] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updated VIF entry in instance network info cache for port 22f6c5a6-c617-44cd-9f61-61773c5a18f6. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 947.549227] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updating instance_info_cache with network_info: [{"id": "f795ad8a-6471-483d-a6b9-a381bad1c376", "address": "fa:16:3e:38:9b:0a", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf795ad8a-64", "ovs_interfaceid": "f795ad8a-6471-483d-a6b9-a381bad1c376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "address": "fa:16:3e:7f:6a:c0", "network": {"id": "69f9d02f-97b1-4b21-80da-80e584d26f8b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1353042792", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.32", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22f6c5a6-c6", "ovs_interfaceid": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "address": "fa:16:3e:8e:31:38", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9e89f7-41", "ovs_interfaceid": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 947.631160] env[65788]: DEBUG nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Received event network-vif-plugged-b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 947.631160] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.631160] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.631160] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.631160] env[65788]: DEBUG nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] No waiting events found dispatching network-vif-plugged-b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 947.631160] env[65788]: WARNING nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Received unexpected event network-vif-plugged-b62b03bf-7b53-4c01-880d-43cd9bf586f6 for instance with vm_state building and task_state spawning. [ 947.631160] env[65788]: DEBUG nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Received event network-changed-b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 947.631160] env[65788]: DEBUG nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Refreshing instance network info cache due to event network-changed-b62b03bf-7b53-4c01-880d-43cd9bf586f6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 947.631503] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Acquiring lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.631503] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Acquired lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.631503] env[65788]: DEBUG nova.network.neutron [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Refreshing network info cache for port b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 947.666414] env[65788]: INFO nova.compute.resource_tracker [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating resource usage from migration 8b1f0895-2e84-4d39-9589-aa78a5a0a9b1 [ 947.683608] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "refresh_cache-c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.683809] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquired lock "refresh_cache-c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.684463] env[65788]: DEBUG nova.network.neutron [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 947.733460] env[65788]: INFO nova.compute.manager [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Took 27.09 seconds to build instance. [ 947.806998] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.750512} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.807482] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8/6a69d645-f89d-43b6-8bdc-b12ce0d88ca8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.807482] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.808059] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8931de13-0c71-4bce-a29a-2bbd2df34f0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.816495] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 947.816495] env[65788]: value = "task-4662709" [ 947.816495] env[65788]: _type = "Task" [ 947.816495] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.826290] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.856553] env[65788]: DEBUG oslo_concurrency.lockutils [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 947.856786] env[65788]: DEBUG nova.compute.manager [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 947.858056] env[65788]: DEBUG nova.compute.manager [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing instance network info cache due to event network-changed-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 947.858056] env[65788]: DEBUG oslo_concurrency.lockutils [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Acquiring lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.858056] env[65788]: DEBUG oslo_concurrency.lockutils [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Acquired lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.858056] env[65788]: DEBUG nova.network.neutron [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Refreshing network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 947.951187] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662708, 'name': Rename_Task, 'duration_secs': 0.181908} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.951482] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.951739] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a23c10a6-eee9-49b4-a42a-6ffbf7ae368b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.959665] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 947.959665] env[65788]: value = "task-4662710" [ 947.959665] env[65788]: _type = "Task" [ 947.959665] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.819633] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Releasing lock "refresh_cache-e0f7c849-315c-4247-a840-d388d48746b9" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.819947] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 948.820170] env[65788]: DEBUG nova.compute.manager [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing instance network info cache due to event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 948.820352] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.820630] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.820717] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 948.825020] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Releasing lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.825020] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-vif-plugged-0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 948.825020] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquiring lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.825020] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.825020] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.825020] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] No waiting events found dispatching network-vif-plugged-0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 948.825020] env[65788]: WARNING nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received unexpected event network-vif-plugged-0c9e89f7-4194-4cd0-86aa-72b4f8655f1e for instance with vm_state building and task_state spawning. [ 948.825020] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-changed-0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 948.825020] env[65788]: DEBUG nova.compute.manager [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Refreshing instance network info cache due to event network-changed-0c9e89f7-4194-4cd0-86aa-72b4f8655f1e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 948.825020] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquiring lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.825020] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Acquired lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.825020] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Refreshing network info cache for port 0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 948.825637] env[65788]: WARNING neutronclient.v2_0.client [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.825731] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.826063] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.836795] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.837256] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.843862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1c932040-7da4-4f93-8303-722ab7a069b7 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.208s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.844456] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.844871] env[65788]: WARNING neutronclient.v2_0.client [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.845698] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.845800] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.853112] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.853112] env[65788]: WARNING oslo_vmware.common.loopingcall [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] task run outlasted interval by 0.39281699999999997 sec [ 948.854637] env[65788]: WARNING neutronclient.v2_0.client [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 948.855326] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 948.855691] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 948.876086] env[65788]: DEBUG nova.compute.manager [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Received event network-vif-plugged-6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 948.876086] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Acquiring lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 948.876086] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 948.876086] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 948.876086] env[65788]: DEBUG nova.compute.manager [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] No waiting events found dispatching network-vif-plugged-6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 948.876086] env[65788]: WARNING nova.compute.manager [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Received unexpected event network-vif-plugged-6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 for instance with vm_state building and task_state spawning. [ 948.876086] env[65788]: DEBUG nova.compute.manager [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Received event network-changed-6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 948.876086] env[65788]: DEBUG nova.compute.manager [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Refreshing instance network info cache due to event network-changed-6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 948.876086] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Acquiring lock "refresh_cache-c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.878665] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-261b1c78-bd39-4486-a613-0f8e42ee3e9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.896553] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523e1eed-5044-5987-ca58-01f16f9d7ecd, 'name': SearchDatastore_Task, 'duration_secs': 0.028732} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.900591] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.900853] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.901094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.901232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 948.901405] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.902162] env[65788]: DEBUG oslo_vmware.api [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 948.902162] env[65788]: value = "task-4662711" [ 948.902162] env[65788]: _type = "Task" [ 948.902162] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.904576] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075805} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.904576] env[65788]: DEBUG oslo_vmware.api [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662710, 'name': PowerOnVM_Task, 'duration_secs': 0.670871} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.906402] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-adc50954-dc0e-45de-9dbc-99ca9d9e35a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.911021] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.911021] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 948.911021] env[65788]: INFO nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Took 9.55 seconds to spawn the instance on the hypervisor. [ 948.911021] env[65788]: DEBUG nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 948.913741] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b807fa0-ed9a-4fae-8a00-f8a731142976 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.916907] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fe6322-d2c1-4157-b512-1cd97ab4ded0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.931419] env[65788]: DEBUG oslo_vmware.api [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662711, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.951117] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.951314] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.961114] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8/6a69d645-f89d-43b6-8bdc-b12ce0d88ca8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.963474] env[65788]: DEBUG nova.network.neutron [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 948.965305] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbe10e86-bc4d-44a8-8a4f-25bdf15911fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.968050] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e977de9-b5aa-4d27-9b06-c556607d151a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.983077] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c97df4c-9eee-4de4-a5ed-73db2c3d7c9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.996008] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 948.996008] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5252fba8-b558-1a8b-a0b3-ee0d2c0aa56a" [ 948.996008] env[65788]: _type = "Task" [ 948.996008] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.998513] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c712a7-3832-4ecb-85a9-73eebb42763f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.004648] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 949.004648] env[65788]: value = "task-4662712" [ 949.004648] env[65788]: _type = "Task" [ 949.004648] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.043950] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1268ac41-5128-438c-928a-eb5c726a8248 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.054700] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.055150] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5252fba8-b558-1a8b-a0b3-ee0d2c0aa56a, 'name': SearchDatastore_Task, 'duration_secs': 0.012414} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.056603] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a058d0dc-60fc-4d97-9c18-5b07f47cab90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.063300] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fe9989-8ae9-4686-b622-a46b757d8928 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.074079] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 949.074079] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b464b-33f5-32e4-5215-5e1cd87e7b3e" [ 949.074079] env[65788]: _type = "Task" [ 949.074079] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.086291] env[65788]: DEBUG nova.compute.provider_tree [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.094560] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b464b-33f5-32e4-5215-5e1cd87e7b3e, 'name': SearchDatastore_Task, 'duration_secs': 0.012616} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.094849] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.095127] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] ec7f7a46-8dd2-4b75-866f-20e73907f1cd/ec7f7a46-8dd2-4b75-866f-20e73907f1cd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.095413] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6a57d47-e0d5-4391-8c5e-ac9f01e35dec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.104436] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 949.104436] env[65788]: value = "task-4662713" [ 949.104436] env[65788]: _type = "Task" [ 949.104436] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.115482] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662713, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.152704] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.153240] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.209728] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.209728] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.288184] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.288797] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.297495] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.298073] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.322463] env[65788]: WARNING neutronclient.v2_0.client [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.323240] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.323604] env[65788]: WARNING openstack [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.354053] env[65788]: WARNING neutronclient.v2_0.client [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.354756] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.355145] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.381834] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "b4f157a7-350c-4fff-8509-44426714846c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.382205] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "b4f157a7-350c-4fff-8509-44426714846c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.382425] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "b4f157a7-350c-4fff-8509-44426714846c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.382811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "b4f157a7-350c-4fff-8509-44426714846c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.382811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "b4f157a7-350c-4fff-8509-44426714846c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.384585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.384783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.384998] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.385170] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.385328] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.389036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.389253] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.390801] env[65788]: INFO nova.compute.manager [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Terminating instance [ 949.392365] env[65788]: INFO nova.compute.manager [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Terminating instance [ 949.394883] env[65788]: WARNING neutronclient.v2_0.client [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.395557] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.395914] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.427773] env[65788]: DEBUG oslo_vmware.api [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662711, 'name': PowerOffVM_Task, 'duration_secs': 0.246298} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.428158] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.428404] env[65788]: DEBUG nova.compute.manager [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 949.429363] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874185b6-2b66-4402-92ff-6c650cc9aa2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.475265] env[65788]: INFO nova.compute.manager [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Took 28.28 seconds to build instance. [ 949.491011] env[65788]: WARNING neutronclient.v2_0.client [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.491788] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.492169] env[65788]: WARNING openstack [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.503729] env[65788]: WARNING neutronclient.v2_0.client [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.504435] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.504792] env[65788]: WARNING openstack [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.530995] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662712, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.543596] env[65788]: DEBUG nova.network.neutron [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Updating instance_info_cache with network_info: [{"id": "6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48", "address": "fa:16:3e:cc:ab:a2", "network": {"id": "dd0e8234-8006-4abe-8a4d-aa8edd49742b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-125145946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99b3e41827194188800da4e0544bc19f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e2a5ea0-49", "ovs_interfaceid": "6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.590320] env[65788]: DEBUG nova.scheduler.client.report [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 949.619858] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662713, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.680747] env[65788]: DEBUG nova.network.neutron [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updated VIF entry in instance network info cache for port b62b03bf-7b53-4c01-880d-43cd9bf586f6. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 949.681219] env[65788]: DEBUG nova.network.neutron [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updating instance_info_cache with network_info: [{"id": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "address": "fa:16:3e:37:b2:1d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62b03bf-7b", "ovs_interfaceid": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.718902] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.720321] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.750813] env[65788]: DEBUG nova.network.neutron [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updated VIF entry in instance network info cache for port 27626e1a-51fe-4f5a-9b93-8ab4c1b8b694. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 949.751798] env[65788]: DEBUG nova.network.neutron [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [{"id": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "address": "fa:16:3e:f8:9f:88", "network": {"id": "ce013ecc-9ca9-42db-8817-cf3b92f2ad2b", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-232286906-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.188", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c4cffe8dfc424459a90e316ac4ac905f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ef746c57-cd18-4883-a0e9-c52937aaf41d", "external-id": "nsx-vlan-transportzone-863", "segmentation_id": 863, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27626e1a-51", "ovs_interfaceid": "27626e1a-51fe-4f5a-9b93-8ab4c1b8b694", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.755894] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updated VIF entry in instance network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 949.756249] env[65788]: DEBUG nova.network.neutron [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.820848] env[65788]: WARNING neutronclient.v2_0.client [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 949.822029] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 949.822029] env[65788]: WARNING openstack [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 949.894915] env[65788]: INFO nova.compute.manager [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Detaching volume 0d9578e1-90ca-4aaa-8fac-028b7f80c67a [ 949.908368] env[65788]: DEBUG nova.compute.manager [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 949.908368] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.908604] env[65788]: DEBUG nova.compute.manager [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 949.908604] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.909443] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e778524-2cc4-4bf4-8a8b-baef0c238f8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.913265] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04208255-8d7e-4cc1-a0ce-350158a193c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.921285] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.924031] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-584550a4-a614-4962-840f-3e5a9a2f8ce7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.925398] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.928034] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updated VIF entry in instance network info cache for port 0c9e89f7-4194-4cd0-86aa-72b4f8655f1e. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 949.928432] env[65788]: DEBUG nova.network.neutron [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updating instance_info_cache with network_info: [{"id": "f795ad8a-6471-483d-a6b9-a381bad1c376", "address": "fa:16:3e:38:9b:0a", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf795ad8a-64", "ovs_interfaceid": "f795ad8a-6471-483d-a6b9-a381bad1c376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "address": "fa:16:3e:7f:6a:c0", "network": {"id": "69f9d02f-97b1-4b21-80da-80e584d26f8b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1353042792", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.32", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22f6c5a6-c6", "ovs_interfaceid": "22f6c5a6-c617-44cd-9f61-61773c5a18f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "address": "fa:16:3e:8e:31:38", "network": {"id": "9a146e07-65ad-4940-90b6-fb79e1e7dcf9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1466060219", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e27fd35-1d7b-4358-92d5-4d34da27b992", "external-id": "nsx-vlan-transportzone-355", "segmentation_id": 355, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9e89f7-41", "ovs_interfaceid": "0c9e89f7-4194-4cd0-86aa-72b4f8655f1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 949.929933] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6b661d7-8e9f-4028-b0bb-2e7a0b9dfb1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.938527] env[65788]: DEBUG oslo_vmware.api [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 949.938527] env[65788]: value = "task-4662714" [ 949.938527] env[65788]: _type = "Task" [ 949.938527] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.941580] env[65788]: DEBUG oslo_vmware.api [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 949.941580] env[65788]: value = "task-4662715" [ 949.941580] env[65788]: _type = "Task" [ 949.941580] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.947347] env[65788]: INFO nova.virt.block_device [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Attempting to driver detach volume 0d9578e1-90ca-4aaa-8fac-028b7f80c67a from mountpoint /dev/sdb [ 949.947925] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 949.948158] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 949.955574] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3efc95f-b002-456f-b8ef-07cba726628a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.959838] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5618083e-3dfe-4474-b142-df0967d10e10 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.650s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.971697] env[65788]: DEBUG oslo_vmware.api [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662715, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.972123] env[65788]: DEBUG oslo_vmware.api [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.993054] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56917995-ded9-4b36-8d4d-6e3ee976923d tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "e0f7c849-315c-4247-a840-d388d48746b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.814s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.995013] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3a211c-769d-4437-9c28-55e256263447 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.005739] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741e126b-8fd0-455a-9877-4877357c9008 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.036690] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdce786-2d9a-4065-9184-06b94f47697d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.048448] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Releasing lock "refresh_cache-c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.049087] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Instance network_info: |[{"id": "6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48", "address": "fa:16:3e:cc:ab:a2", "network": {"id": "dd0e8234-8006-4abe-8a4d-aa8edd49742b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-125145946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99b3e41827194188800da4e0544bc19f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e2a5ea0-49", "ovs_interfaceid": "6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 950.049621] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662712, 'name': ReconfigVM_Task, 'duration_secs': 0.592364} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.062215] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Acquired lock "refresh_cache-c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.062737] env[65788]: DEBUG nova.network.neutron [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Refreshing network info cache for port 6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 950.067239] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:ab:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.073993] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Creating folder: Project (99b3e41827194188800da4e0544bc19f). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 950.073993] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8/6a69d645-f89d-43b6-8bdc-b12ce0d88ca8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.074218] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The volume has not been displaced from its original location: [datastore1] volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a/volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 950.080291] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfiguring VM instance instance-00000038 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 950.082245] env[65788]: WARNING neutronclient.v2_0.client [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.084031] env[65788]: WARNING openstack [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.084031] env[65788]: WARNING openstack [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.094037] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2434e3ab-6ffd-4eb2-8e81-f281f22dbac8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.096356] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0448862-2f4f-4bf0-ba51-b704bc097349 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.098162] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15082a8a-f068-4325-a0de-3f5bf37a9b8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.117430] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.464s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.117653] env[65788]: INFO nova.compute.manager [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Migrating [ 950.134261] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.205s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.134261] env[65788]: DEBUG nova.objects.instance [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lazy-loading 'resources' on Instance uuid 37471e59-1809-4df3-8c40-20921d04d18e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.140056] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 950.140401] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 950.158773] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Created folder: Project (99b3e41827194188800da4e0544bc19f) in parent group-v910111. [ 950.158773] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Creating folder: Instances. Parent ref: group-v910343. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 950.159538] env[65788]: DEBUG oslo_vmware.api [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 950.159538] env[65788]: value = "task-4662718" [ 950.159538] env[65788]: _type = "Task" [ 950.159538] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.159798] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 950.159798] env[65788]: value = "task-4662717" [ 950.159798] env[65788]: _type = "Task" [ 950.159798] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.160828] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.161128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.161430] env[65788]: DEBUG nova.network.neutron [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 950.165333] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d3d5aa6-aa58-4e4e-b55e-b23e1b115774 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.186668] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Releasing lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.186925] env[65788]: DEBUG nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 950.187103] env[65788]: DEBUG nova.compute.manager [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing instance network info cache due to event network-changed-891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 950.187290] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Acquiring lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.187618] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662713, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.641272} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.189503] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] ec7f7a46-8dd2-4b75-866f-20e73907f1cd/ec7f7a46-8dd2-4b75-866f-20e73907f1cd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.189785] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.190610] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a65fb575-6cfd-4557-bdf4-5b5f257d5913 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.207584] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662717, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.207902] env[65788]: DEBUG oslo_vmware.api [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662718, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.211501] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Created folder: Instances in parent group-v910343. [ 950.211501] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 950.211501] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 950.211501] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-616e5595-2cb2-4014-8bcc-133c78de20b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.233101] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 950.233101] env[65788]: value = "task-4662720" [ 950.233101] env[65788]: _type = "Task" [ 950.233101] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.243681] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.243681] env[65788]: value = "task-4662721" [ 950.243681] env[65788]: _type = "Task" [ 950.243681] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.247349] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662720, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.259649] env[65788]: DEBUG oslo_concurrency.lockutils [req-b114c0c7-ab36-4457-9212-c379d1b95006 req-9cc82815-81dc-4635-85e6-e422090d85b9 service nova] Releasing lock "refresh_cache-3a06a833-5aaa-4b5d-88b3-8a1d469580af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.260886] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662721, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.260981] env[65788]: DEBUG oslo_concurrency.lockutils [req-188b2e8e-17ce-410d-a257-efaabb283fc7 req-cbe678fb-5e7d-496f-82a3-08915d42e74f service nova] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.265064] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Acquired lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.265340] env[65788]: DEBUG nova.network.neutron [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Refreshing network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 950.292464] env[65788]: WARNING openstack [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.292986] env[65788]: WARNING openstack [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.368824] env[65788]: WARNING neutronclient.v2_0.client [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.369506] env[65788]: WARNING openstack [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.369994] env[65788]: WARNING openstack [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.433742] env[65788]: DEBUG oslo_concurrency.lockutils [req-d545d7dc-acb6-4bcc-951a-0d1f891bd50e req-2fdac815-a05d-4f9b-bd1c-074a1d0cb37c service nova] Releasing lock "refresh_cache-6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.465243] env[65788]: DEBUG oslo_vmware.api [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662715, 'name': PowerOffVM_Task, 'duration_secs': 0.287245} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.465511] env[65788]: DEBUG oslo_vmware.api [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662714, 'name': PowerOffVM_Task, 'duration_secs': 0.25426} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.468769] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.468970] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.469352] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.469466] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 950.469854] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a726db3-cd7a-454b-a9c9-c83214c87da8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.471546] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-35527295-c826-4ba6-968e-86a6d9e8c547 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.493112] env[65788]: DEBUG nova.network.neutron [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Updated VIF entry in instance network info cache for port 6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 950.493580] env[65788]: DEBUG nova.network.neutron [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Updating instance_info_cache with network_info: [{"id": "6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48", "address": "fa:16:3e:cc:ab:a2", "network": {"id": "dd0e8234-8006-4abe-8a4d-aa8edd49742b", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-125145946-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99b3e41827194188800da4e0544bc19f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e2a5ea0-49", "ovs_interfaceid": "6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 950.578017] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.578017] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.578017] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Deleting the datastore file [datastore2] b4f157a7-350c-4fff-8509-44426714846c {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.578017] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27d3cbaf-7d99-4b14-a483-358eb9e1bc3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.584059] env[65788]: DEBUG oslo_vmware.api [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for the task: (returnval){ [ 950.584059] env[65788]: value = "task-4662724" [ 950.584059] env[65788]: _type = "Task" [ 950.584059] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.595954] env[65788]: DEBUG oslo_vmware.api [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.611049] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 950.611258] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 950.611437] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleting the datastore file [datastore2] d2d614f8-15d9-47d6-9e66-8e1f8fcc18da {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.611722] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-425c41d3-32c1-40bd-b27e-52e8cd914b42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.619170] env[65788]: DEBUG oslo_vmware.api [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for the task: (returnval){ [ 950.619170] env[65788]: value = "task-4662725" [ 950.619170] env[65788]: _type = "Task" [ 950.619170] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.623719] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f03d8a-8406-4415-890f-e224d704e442 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.634176] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97f6f27-c301-40bf-b581-f258cc27b979 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.638431] env[65788]: DEBUG oslo_vmware.api [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662725, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.668393] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 950.675340] env[65788]: DEBUG nova.compute.manager [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 950.675340] env[65788]: DEBUG nova.compute.manager [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing instance network info cache due to event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 950.675340] env[65788]: DEBUG oslo_concurrency.lockutils [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.675340] env[65788]: DEBUG oslo_concurrency.lockutils [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.675340] env[65788]: DEBUG nova.network.neutron [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 950.675340] env[65788]: WARNING neutronclient.v2_0.client [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.675340] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.675914] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.683252] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0a3b5e-2f34-4dd7-b87a-cdcaf45ad4b0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.704418] env[65788]: DEBUG oslo_vmware.api [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662718, 'name': ReconfigVM_Task, 'duration_secs': 0.377594} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.705709] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662717, 'name': Rename_Task, 'duration_secs': 0.27553} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.705709] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Reconfigured VM instance instance-00000038 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 950.711358] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67e615f-d8fa-4a0a-947c-1fefb92b4134 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.715274] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.716124] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90544a4b-e55c-48ef-a7e5-ced6856c2d5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.729072] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21d33b7b-60e9-47ad-8d7d-819672b4a1b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.741242] env[65788]: DEBUG nova.compute.provider_tree [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.750711] env[65788]: DEBUG oslo_vmware.api [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 950.750711] env[65788]: value = "task-4662726" [ 950.750711] env[65788]: _type = "Task" [ 950.750711] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.752252] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 950.752252] env[65788]: value = "task-4662727" [ 950.752252] env[65788]: _type = "Task" [ 950.752252] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.768201] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662720, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089808} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.778192] env[65788]: WARNING neutronclient.v2_0.client [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.778999] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.779379] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.786392] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.786760] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662727, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.787083] env[65788]: DEBUG oslo_vmware.api [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662726, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.787658] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662721, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.788992] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6754feb9-2ac2-46ba-a8df-108cc20963ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.817012] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] ec7f7a46-8dd2-4b75-866f-20e73907f1cd/ec7f7a46-8dd2-4b75-866f-20e73907f1cd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.817522] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80701110-050a-444c-9149-13220d82a80f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.836327] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.836729] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.849966] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 950.849966] env[65788]: value = "task-4662728" [ 950.849966] env[65788]: _type = "Task" [ 950.849966] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.863382] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662728, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.887560] env[65788]: DEBUG nova.objects.instance [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.938335] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.938335] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 950.966627] env[65788]: WARNING neutronclient.v2_0.client [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 950.968048] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 950.968048] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.002925] env[65788]: DEBUG oslo_concurrency.lockutils [req-2d7834e4-8643-48d5-bd6f-4aeec0e8216b req-70e7c8aa-229a-4ef2-b2ae-26d1ec46591b service nova] Releasing lock "refresh_cache-c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.103929] env[65788]: DEBUG oslo_vmware.api [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Task: {'id': task-4662724, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.493619} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.104623] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.104728] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.105211] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.105256] env[65788]: INFO nova.compute.manager [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] [instance: b4f157a7-350c-4fff-8509-44426714846c] Took 1.20 seconds to destroy the instance on the hypervisor. [ 951.105586] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 951.105882] env[65788]: DEBUG nova.compute.manager [-] [instance: b4f157a7-350c-4fff-8509-44426714846c] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 951.106022] env[65788]: DEBUG nova.network.neutron [-] [instance: b4f157a7-350c-4fff-8509-44426714846c] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 951.106363] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.107148] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.107546] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.128750] env[65788]: WARNING neutronclient.v2_0.client [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.129464] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.129862] env[65788]: WARNING openstack [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.143786] env[65788]: DEBUG oslo_vmware.api [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Task: {'id': task-4662725, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.394193} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.144271] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.144647] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.144736] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.144827] env[65788]: INFO nova.compute.manager [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Took 1.24 seconds to destroy the instance on the hypervisor. [ 951.145117] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 951.145324] env[65788]: DEBUG nova.compute.manager [-] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 951.145419] env[65788]: DEBUG nova.network.neutron [-] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 951.145804] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.146188] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.146414] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.165364] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.187909] env[65788]: WARNING neutronclient.v2_0.client [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.191539] env[65788]: WARNING openstack [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.191539] env[65788]: WARNING openstack [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.197363] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.201302] env[65788]: DEBUG nova.network.neutron [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance_info_cache with network_info: [{"id": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "address": "fa:16:3e:0a:42:08", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2184dd-c8", "ovs_interfaceid": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.249151] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.255496] env[65788]: DEBUG nova.scheduler.client.report [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 951.274635] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662721, 'name': CreateVM_Task, 'duration_secs': 0.816261} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.274635] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 951.274635] env[65788]: WARNING neutronclient.v2_0.client [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.274841] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.275118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.275527] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 951.279591] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5063dd9-89f2-47f3-89df-45da6567020a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.282490] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662727, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.286836] env[65788]: DEBUG oslo_vmware.api [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662726, 'name': ReconfigVM_Task, 'duration_secs': 0.233775} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.290026] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910310', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'name': 'volume-0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f160e4a5-1e91-495e-800e-116ec435d8e1', 'attached_at': '', 'detached_at': '', 'volume_id': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a', 'serial': '0d9578e1-90ca-4aaa-8fac-028b7f80c67a'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 951.292353] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 951.292353] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522735cc-bd9d-4709-8f7e-a6a436a65626" [ 951.292353] env[65788]: _type = "Task" [ 951.292353] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.307163] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522735cc-bd9d-4709-8f7e-a6a436a65626, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.365623] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662728, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.369354] env[65788]: DEBUG nova.network.neutron [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updated VIF entry in instance network info cache for port 891ac553-14f8-41d2-a6e4-c24ef3c62c6a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 951.369880] env[65788]: DEBUG nova.network.neutron [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [{"id": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "address": "fa:16:3e:58:21:32", "network": {"id": "306b88b5-be8e-4339-88ed-5c8b6ccca839", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1402321202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "483bc5fc16c443ab89591367b0a588ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891ac553-14", "ovs_interfaceid": "891ac553-14f8-41d2-a6e4-c24ef3c62c6a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.393288] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.393600] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.393909] env[65788]: DEBUG nova.network.neutron [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 951.394206] env[65788]: DEBUG nova.objects.instance [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'info_cache' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.559935] env[65788]: WARNING openstack [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.562425] env[65788]: WARNING openstack [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.655652] env[65788]: WARNING neutronclient.v2_0.client [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 951.656370] env[65788]: WARNING openstack [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 951.656724] env[65788]: WARNING openstack [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 951.705146] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.771024] env[65788]: DEBUG nova.network.neutron [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updated VIF entry in instance network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 951.771024] env[65788]: DEBUG nova.network.neutron [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 951.776765] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.781547] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.476s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.783330] env[65788]: INFO nova.compute.claims [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.787050] env[65788]: DEBUG oslo_vmware.api [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662727, 'name': PowerOnVM_Task, 'duration_secs': 0.805546} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.787050] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.787050] env[65788]: INFO nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Took 17.02 seconds to spawn the instance on the hypervisor. [ 951.787050] env[65788]: DEBUG nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 951.787733] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3d668a-fc2a-4bc0-b018-8189df50c9c1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.809134] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522735cc-bd9d-4709-8f7e-a6a436a65626, 'name': SearchDatastore_Task, 'duration_secs': 0.013946} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.809437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.809673] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.809897] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.810083] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.810252] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.810537] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2207be0f-40b4-4e4f-a006-8f21722677fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.824782] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.824782] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 951.824782] env[65788]: INFO nova.scheduler.client.report [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted allocations for instance 37471e59-1809-4df3-8c40-20921d04d18e [ 951.825118] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-772bfeb7-d99c-415e-bfb1-e354a6d565e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.837062] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 951.837062] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ef8601-011d-2217-8081-d1bd485efd05" [ 951.837062] env[65788]: _type = "Task" [ 951.837062] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.849864] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ef8601-011d-2217-8081-d1bd485efd05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.866473] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662728, 'name': ReconfigVM_Task, 'duration_secs': 0.528139} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.867913] env[65788]: DEBUG nova.objects.instance [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'flavor' on Instance uuid f160e4a5-1e91-495e-800e-116ec435d8e1 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.868799] env[65788]: WARNING oslo_messaging._drivers.amqpdriver [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 951.869850] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Reconfigured VM instance instance-0000004d to attach disk [datastore1] ec7f7a46-8dd2-4b75-866f-20e73907f1cd/ec7f7a46-8dd2-4b75-866f-20e73907f1cd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.871560] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e87a7849-999e-421b-9f37-b548425e80dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.873910] env[65788]: DEBUG oslo_concurrency.lockutils [req-f2202390-3cdb-4aca-bd7a-a32b44827b73 req-81f022e4-0f3a-4307-b830-8460976edbe1 service nova] Releasing lock "refresh_cache-b4f157a7-350c-4fff-8509-44426714846c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.880388] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 951.880388] env[65788]: value = "task-4662729" [ 951.880388] env[65788]: _type = "Task" [ 951.880388] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.890318] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662729, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.898301] env[65788]: DEBUG nova.objects.base [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 951.960978] env[65788]: DEBUG nova.network.neutron [-] [instance: b4f157a7-350c-4fff-8509-44426714846c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.097193] env[65788]: DEBUG nova.network.neutron [-] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.204647] env[65788]: DEBUG nova.compute.manager [req-002ca7f8-4115-4d0b-b34b-af070eabd251 req-de1ae96e-7d2d-4dfa-95d2-677d199e9c9f service nova] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Received event network-vif-deleted-3a1140bb-e791-40d3-a08e-d01e1bfb79b6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 952.273479] env[65788]: DEBUG oslo_concurrency.lockutils [req-8a38ead0-c7e5-4ea9-a95a-605a3583d7b7 req-08238397-73dd-4c17-9c4a-23576ff4a3a7 service nova] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.311623] env[65788]: INFO nova.compute.manager [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Took 32.28 seconds to build instance. [ 952.335752] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52feeda4-4bd6-4463-ae7b-033d6f1d0903 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "37471e59-1809-4df3-8c40-20921d04d18e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.165s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.348238] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ef8601-011d-2217-8081-d1bd485efd05, 'name': SearchDatastore_Task, 'duration_secs': 0.023345} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.349762] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3acc9b38-f0bd-4521-a7d4-3bb3aaa9bd33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.359603] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 952.359603] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52069a7f-28a0-1ae3-e47c-9398d10ca0bb" [ 952.359603] env[65788]: _type = "Task" [ 952.359603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.368541] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52069a7f-28a0-1ae3-e47c-9398d10ca0bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.393345] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662729, 'name': Rename_Task, 'duration_secs': 0.237987} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.393669] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.393926] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e06e391a-2ae7-46e8-9278-ef09daf32d1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.401963] env[65788]: WARNING neutronclient.v2_0.client [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.402639] env[65788]: WARNING openstack [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.403066] env[65788]: WARNING openstack [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.409645] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 952.409645] env[65788]: value = "task-4662730" [ 952.409645] env[65788]: _type = "Task" [ 952.409645] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.419342] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662730, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.463790] env[65788]: INFO nova.compute.manager [-] [instance: b4f157a7-350c-4fff-8509-44426714846c] Took 1.36 seconds to deallocate network for instance. [ 952.522302] env[65788]: WARNING openstack [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.522933] env[65788]: WARNING openstack [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.590576] env[65788]: WARNING neutronclient.v2_0.client [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 952.591371] env[65788]: WARNING openstack [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 952.591810] env[65788]: WARNING openstack [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 952.603120] env[65788]: INFO nova.compute.manager [-] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Took 1.46 seconds to deallocate network for instance. [ 952.680742] env[65788]: DEBUG nova.network.neutron [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [{"id": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "address": "fa:16:3e:88:8c:66", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ac3e7fa-b2", "ovs_interfaceid": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 952.798252] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "af9120a6-2cea-46da-ba06-6036b9beda78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 952.798496] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 952.821641] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8b01e04-50dd-4f8c-a192-6f3001e95dbb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.804s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.833214] env[65788]: DEBUG nova.compute.manager [req-6c8e11b3-c0a7-4f7a-a414-b0969614a064 req-f1d1f291-9555-450e-a805-730205c06488 service nova] [instance: b4f157a7-350c-4fff-8509-44426714846c] Received event network-vif-deleted-891ac553-14f8-41d2-a6e4-c24ef3c62c6a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 952.857078] env[65788]: DEBUG nova.compute.manager [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 952.858231] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bde606f-cecb-415f-ad58-2aca95f289e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.872830] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52069a7f-28a0-1ae3-e47c-9398d10ca0bb, 'name': SearchDatastore_Task, 'duration_secs': 0.02201} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.876308] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 952.876689] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb/c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 952.877823] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39945e91-94f1-477d-aabb-c0489fa5b1f6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.880871] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0bd89420-bddc-4376-862b-457b0be269e6 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.491s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.886958] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 952.886958] env[65788]: value = "task-4662731" [ 952.886958] env[65788]: _type = "Task" [ 952.886958] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.897298] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662731, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.921433] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662730, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.974019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.111149] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.185179] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.190813] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f265b5a-da07-4324-97e0-df0d7b0195cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.207709] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f2527b-652f-4afa-87c9-7cc877529bd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.256812] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6d3808-2969-4be7-a040-6e0e98b38e4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.260432] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ccffd0-e75f-4b94-ba8e-fdc7eecb79ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.283293] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11037c2-c3d2-488a-be7f-c0d2ecf546ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.287758] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 953.305776] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 953.309689] env[65788]: DEBUG nova.compute.provider_tree [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.378143] env[65788]: INFO nova.compute.manager [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] instance snapshotting [ 953.381792] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c00306-1844-4219-9c30-5730c3981fef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.410214] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c49645e-a15d-4782-806c-4b4d71463151 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.422128] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662731, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.431147] env[65788]: DEBUG oslo_vmware.api [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662730, 'name': PowerOnVM_Task, 'duration_secs': 0.843152} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.432011] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.432335] env[65788]: INFO nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Took 9.60 seconds to spawn the instance on the hypervisor. [ 953.432526] env[65788]: DEBUG nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 953.433533] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67a0b1f-a344-4ca2-8471-1e72b8711893 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.495122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.495122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.495122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.495299] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.495515] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.497872] env[65788]: INFO nova.compute.manager [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Terminating instance [ 953.794218] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 953.794558] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4573de6e-ffce-4f40-a377-1fdcc3db6565 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.803585] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 953.803585] env[65788]: value = "task-4662732" [ 953.803585] env[65788]: _type = "Task" [ 953.803585] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.815911] env[65788]: DEBUG nova.scheduler.client.report [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 953.819595] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.842240] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.902697] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662731, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687591} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.902992] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb/c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.903265] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.903539] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbfc5310-c59f-48b2-b093-0baf41ec8ec5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.911959] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 953.911959] env[65788]: value = "task-4662733" [ 953.911959] env[65788]: _type = "Task" [ 953.911959] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.923853] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662733, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.933348] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 953.933693] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-365db8b8-beb1-4e2e-9e8f-923b24682e9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.943279] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 953.943279] env[65788]: value = "task-4662734" [ 953.943279] env[65788]: _type = "Task" [ 953.943279] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.963033] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662734, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.963207] env[65788]: INFO nova.compute.manager [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Took 22.78 seconds to build instance. [ 954.001619] env[65788]: DEBUG nova.compute.manager [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 954.001906] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.003055] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f2ca5b-a242-4864-9d98-72b7cc73798a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.013234] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.013625] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1be03449-8f8e-4e50-a472-83bcafdabdfc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.020727] env[65788]: DEBUG oslo_vmware.api [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 954.020727] env[65788]: value = "task-4662735" [ 954.020727] env[65788]: _type = "Task" [ 954.020727] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.032548] env[65788]: DEBUG oslo_vmware.api [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.037323] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.037629] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.037860] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "f160e4a5-1e91-495e-800e-116ec435d8e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.038100] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.038310] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.041239] env[65788]: INFO nova.compute.manager [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Terminating instance [ 954.199591] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.199954] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a1adb80-da2c-4b31-ad7b-38d1936b1808 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.208716] env[65788]: DEBUG oslo_vmware.api [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 954.208716] env[65788]: value = "task-4662736" [ 954.208716] env[65788]: _type = "Task" [ 954.208716] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.219346] env[65788]: DEBUG oslo_vmware.api [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662736, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.315761] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662732, 'name': PowerOffVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.321570] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.322168] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 954.328722] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.822s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.330260] env[65788]: INFO nova.compute.claims [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.429098] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662733, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072787} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.429422] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.430538] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d62cca4-aa69-4c41-9e35-14c3490eec80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.461147] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb/c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.464580] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5d30617-87f5-4ef5-a5ec-53eae0bc9eac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.480822] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a3237374-c546-46a2-86cd-b1b1ee67a68b tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.306s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.489057] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662734, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.490820] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 954.490820] env[65788]: value = "task-4662737" [ 954.490820] env[65788]: _type = "Task" [ 954.490820] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.500110] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662737, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.532396] env[65788]: DEBUG oslo_vmware.api [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662735, 'name': PowerOffVM_Task, 'duration_secs': 0.477044} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.532749] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.532942] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.533307] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1a6c3e6-3355-4f1d-b10e-8c6b6f9d0cbc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.545943] env[65788]: DEBUG nova.compute.manager [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 954.546230] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.547493] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94aa5acc-371f-44e8-aa0c-3826f4e00c02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.556552] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.557082] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2debc3d9-0372-476a-abc4-9fe72f2dfc0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.566834] env[65788]: DEBUG oslo_vmware.api [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 954.566834] env[65788]: value = "task-4662739" [ 954.566834] env[65788]: _type = "Task" [ 954.566834] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.579797] env[65788]: DEBUG oslo_vmware.api [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.720519] env[65788]: DEBUG oslo_vmware.api [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662736, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.721992] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.722258] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.722428] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Deleting the datastore file [datastore1] 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.722686] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c337621-5db9-491b-bf1f-58f3868859b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.732183] env[65788]: DEBUG oslo_vmware.api [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 954.732183] env[65788]: value = "task-4662740" [ 954.732183] env[65788]: _type = "Task" [ 954.732183] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.740699] env[65788]: DEBUG oslo_vmware.api [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.817736] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662732, 'name': PowerOffVM_Task, 'duration_secs': 0.689363} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.818767] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.818767] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 954.849189] env[65788]: DEBUG nova.compute.utils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 954.852195] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 954.852406] env[65788]: DEBUG nova.network.neutron [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 954.852958] env[65788]: WARNING neutronclient.v2_0.client [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.853228] env[65788]: WARNING neutronclient.v2_0.client [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 954.853866] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 954.854280] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 954.915750] env[65788]: DEBUG nova.policy [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88bdb2dca0ab439e9d0ffd7a063986c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b693f0eeee664f1ab23b73bb1ea6da0a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 954.968113] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662734, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.008446] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662737, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.013759] env[65788]: DEBUG nova.compute.manager [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Received event network-changed-b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 955.013971] env[65788]: DEBUG nova.compute.manager [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Refreshing instance network info cache due to event network-changed-b62b03bf-7b53-4c01-880d-43cd9bf586f6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 955.014304] env[65788]: DEBUG oslo_concurrency.lockutils [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Acquiring lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.015072] env[65788]: DEBUG oslo_concurrency.lockutils [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Acquired lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 955.015072] env[65788]: DEBUG nova.network.neutron [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Refreshing network info cache for port b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 955.077983] env[65788]: DEBUG oslo_vmware.api [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662739, 'name': PowerOffVM_Task, 'duration_secs': 0.262594} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.078661] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.078661] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.078820] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b849b34e-43ca-4fc1-b4c4-a0078942faae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.159780] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.159959] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.160155] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore2] f160e4a5-1e91-495e-800e-116ec435d8e1 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.160394] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0ff65ae-934e-4621-93fb-33f2edcf73b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.170428] env[65788]: DEBUG oslo_vmware.api [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 955.170428] env[65788]: value = "task-4662742" [ 955.170428] env[65788]: _type = "Task" [ 955.170428] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.180881] env[65788]: DEBUG oslo_vmware.api [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662742, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.221447] env[65788]: DEBUG oslo_vmware.api [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662736, 'name': PowerOnVM_Task, 'duration_secs': 0.519032} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.221826] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.222025] env[65788]: DEBUG nova.compute.manager [None req-e2589638-737a-4abe-b93c-e7afcf598d07 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 955.224829] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b7adaf1-60a5-4681-a5fa-005c159c0b84 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.246763] env[65788]: DEBUG oslo_vmware.api [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340279} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.246763] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.246763] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.246763] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.246763] env[65788]: INFO nova.compute.manager [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Took 1.24 seconds to destroy the instance on the hypervisor. [ 955.246763] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 955.246763] env[65788]: DEBUG nova.compute.manager [-] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 955.246763] env[65788]: DEBUG nova.network.neutron [-] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 955.246763] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.246763] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.247494] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.281837] env[65788]: DEBUG nova.network.neutron [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Successfully created port: a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 955.316896] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.326492] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 955.326762] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 955.326923] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 955.327132] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 955.327298] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 955.327450] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 955.327657] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.327831] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 955.328357] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 955.328577] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 955.328804] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 955.334256] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83c3bf82-87ca-4586-a6ae-63a34829db27 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.351972] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 955.351972] env[65788]: value = "task-4662743" [ 955.351972] env[65788]: _type = "Task" [ 955.351972] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.366731] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 955.373799] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662743, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.469992] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662734, 'name': CreateSnapshot_Task, 'duration_secs': 1.300252} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.474564] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 955.476226] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbece12d-d069-43e9-87e1-92b2d8d25266 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.508881] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662737, 'name': ReconfigVM_Task, 'duration_secs': 0.747157} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.509159] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Reconfigured VM instance instance-0000004e to attach disk [datastore2] c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb/c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.512221] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5ee43ef-d486-41a5-8c34-9ede3318bdf2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.521821] env[65788]: WARNING neutronclient.v2_0.client [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.522454] env[65788]: WARNING openstack [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.524313] env[65788]: WARNING openstack [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.536571] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 955.536571] env[65788]: value = "task-4662744" [ 955.536571] env[65788]: _type = "Task" [ 955.536571] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.548488] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662744, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.686594] env[65788]: DEBUG oslo_vmware.api [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662742, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.285303} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.689938] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.689938] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.689938] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.689938] env[65788]: INFO nova.compute.manager [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 955.689938] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 955.689938] env[65788]: DEBUG nova.compute.manager [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 955.689938] env[65788]: DEBUG nova.network.neutron [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 955.689938] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 955.689938] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 955.689938] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 955.808419] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f712de52-6519-4bc1-a386-9da9338e833e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.819911] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024f4393-7dea-45a2-b8c7-cd539d359eb5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.853506] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e61f014-b405-4450-8e03-b8e34f46747a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.868017] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c389cb72-e76d-4536-8ce9-361784466603 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.872276] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662743, 'name': ReconfigVM_Task, 'duration_secs': 0.246228} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.875784] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 955.894427] env[65788]: DEBUG nova.compute.provider_tree [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.004627] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 956.005929] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 956.008180] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0c1f9657-18ff-4c3e-ae4a-8c0b43f00bdb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.017712] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 956.017712] env[65788]: value = "task-4662745" [ 956.017712] env[65788]: _type = "Task" [ 956.017712] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.027728] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662745, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.050212] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662744, 'name': Rename_Task, 'duration_secs': 0.316617} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.050711] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.050827] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2857f69e-c9fe-4e57-850e-6d17f7c5dc17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.063646] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 956.063646] env[65788]: value = "task-4662746" [ 956.063646] env[65788]: _type = "Task" [ 956.063646] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.074308] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.383415] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 956.387045] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 956.387327] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 956.387485] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 956.387672] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 956.387863] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 956.388039] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 956.388258] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.388417] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 956.388629] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 956.388824] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 956.389011] env[65788]: DEBUG nova.virt.hardware [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 956.394689] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Reconfiguring VM instance instance-00000049 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 956.395089] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60fb7fd7-431d-4fa4-9b13-7dc0441eaab3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.411695] env[65788]: DEBUG nova.scheduler.client.report [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 956.421786] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 956.422065] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 956.422230] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 956.422453] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 956.422645] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 956.422733] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 956.422953] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.423128] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 956.423299] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 956.423466] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 956.423663] env[65788]: DEBUG nova.virt.hardware [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 956.424590] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff84a81a-d512-45d5-a468-c60d23e3f14e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.428675] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 956.428675] env[65788]: value = "task-4662747" [ 956.428675] env[65788]: _type = "Task" [ 956.428675] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.436772] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff91676f-489d-4204-bf48-ce9a4a188119 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.444472] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662747, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.529353] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662745, 'name': CloneVM_Task} progress is 93%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.575017] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662746, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.776022] env[65788]: DEBUG nova.network.neutron [-] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 956.853018] env[65788]: WARNING openstack [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 956.853767] env[65788]: WARNING openstack [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 956.895516] env[65788]: DEBUG nova.compute.manager [req-1bbb8e2e-874d-4c94-918b-56c63c4820e3 req-0b16a16f-9315-4696-a489-25af53d71ce3 service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-vif-deleted-22f6c5a6-c617-44cd-9f61-61773c5a18f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 956.920022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.920022] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 956.925858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.387s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.943905] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662747, 'name': ReconfigVM_Task, 'duration_secs': 0.318493} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.944416] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Reconfigured VM instance instance-00000049 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 956.945456] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f088b1e1-028e-4eba-9e3c-5ff5d1a6a899 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.986776] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e/2cb53fe5-aa58-479e-9090-0d8509cf164e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.987161] env[65788]: DEBUG nova.network.neutron [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Successfully updated port: a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 956.989296] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86eb23cf-86b1-4069-a661-01298ed9876b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.007813] env[65788]: WARNING neutronclient.v2_0.client [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.008465] env[65788]: WARNING openstack [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.009583] env[65788]: WARNING openstack [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.027366] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 957.027366] env[65788]: value = "task-4662748" [ 957.027366] env[65788]: _type = "Task" [ 957.027366] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.038942] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662745, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.047287] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662748, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.078529] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662746, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.175403] env[65788]: DEBUG nova.network.neutron [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updated VIF entry in instance network info cache for port b62b03bf-7b53-4c01-880d-43cd9bf586f6. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 957.175914] env[65788]: DEBUG nova.network.neutron [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updating instance_info_cache with network_info: [{"id": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "address": "fa:16:3e:37:b2:1d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb62b03bf-7b", "ovs_interfaceid": "b62b03bf-7b53-4c01-880d-43cd9bf586f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.277446] env[65788]: DEBUG nova.network.neutron [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 957.280085] env[65788]: INFO nova.compute.manager [-] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Took 2.03 seconds to deallocate network for instance. [ 957.429321] env[65788]: DEBUG nova.compute.utils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 957.444457] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 957.444873] env[65788]: DEBUG nova.network.neutron [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 957.445480] env[65788]: WARNING neutronclient.v2_0.client [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.446875] env[65788]: WARNING neutronclient.v2_0.client [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.449142] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.449629] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.464045] env[65788]: DEBUG nova.compute.manager [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Received event network-vif-plugged-a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 957.464327] env[65788]: DEBUG oslo_concurrency.lockutils [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Acquiring lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.464546] env[65788]: DEBUG oslo_concurrency.lockutils [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.464712] env[65788]: DEBUG oslo_concurrency.lockutils [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.464880] env[65788]: DEBUG nova.compute.manager [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] No waiting events found dispatching network-vif-plugged-a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 957.465100] env[65788]: WARNING nova.compute.manager [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Received unexpected event network-vif-plugged-a0e489a6-9195-4339-a917-3486d9f0da45 for instance with vm_state building and task_state spawning. [ 957.465344] env[65788]: DEBUG nova.compute.manager [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Received event network-changed-a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 957.465513] env[65788]: DEBUG nova.compute.manager [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Refreshing instance network info cache due to event network-changed-a0e489a6-9195-4339-a917-3486d9f0da45. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 957.465707] env[65788]: DEBUG oslo_concurrency.lockutils [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Acquiring lock "refresh_cache-a79bbb7e-7b78-4900-9b83-bfce94bb41bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.465838] env[65788]: DEBUG oslo_concurrency.lockutils [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Acquired lock "refresh_cache-a79bbb7e-7b78-4900-9b83-bfce94bb41bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 957.465988] env[65788]: DEBUG nova.network.neutron [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Refreshing network info cache for port a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 957.511591] env[65788]: DEBUG nova.policy [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a1e3cda6bf9439982465aec80977cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e1713c7002a413fb27180469fded83e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 957.514772] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "refresh_cache-a79bbb7e-7b78-4900-9b83-bfce94bb41bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.537107] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662745, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.547823] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662748, 'name': ReconfigVM_Task, 'duration_secs': 0.389151} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.548148] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e/2cb53fe5-aa58-479e-9090-0d8509cf164e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.548394] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 957.576320] env[65788]: DEBUG oslo_vmware.api [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662746, 'name': PowerOnVM_Task, 'duration_secs': 1.490414} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.576713] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.576893] env[65788]: INFO nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Took 11.09 seconds to spawn the instance on the hypervisor. [ 957.577148] env[65788]: DEBUG nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 957.578501] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f5ba50-f6b1-4309-92c4-e9749e459801 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.679464] env[65788]: DEBUG oslo_concurrency.lockutils [req-0ab75a6e-0a5f-4ece-bc60-f0d8c69c5f96 req-98019b7b-962b-4dd2-9c00-900ec221248c service nova] Releasing lock "refresh_cache-ec7f7a46-8dd2-4b75-866f-20e73907f1cd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 957.784260] env[65788]: INFO nova.compute.manager [-] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Took 2.09 seconds to deallocate network for instance. [ 957.790253] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.808857] env[65788]: DEBUG nova.network.neutron [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Successfully created port: f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 957.945021] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 957.960896] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Applying migration context for instance 2cb53fe5-aa58-479e-9090-0d8509cf164e as it has an incoming, in-progress migration 8b1f0895-2e84-4d39-9589-aa78a5a0a9b1. Migration status is migrating {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 957.963513] env[65788]: INFO nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating resource usage from migration 8b1f0895-2e84-4d39-9589-aa78a5a0a9b1 [ 957.969700] env[65788]: WARNING neutronclient.v2_0.client [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 957.970425] env[65788]: WARNING openstack [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 957.970818] env[65788]: WARNING openstack [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 957.986095] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.986259] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6215ea56-f99c-42f3-a2d5-f4d31952c8e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.986375] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9438ab56-1b4c-4778-a608-de319ab0ee43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987201] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d78d36e6-a542-4ba8-9e29-b8a7b28c559e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987201] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f160e4a5-1e91-495e-800e-116ec435d8e1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987201] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance a35ced42-4317-49b4-b4cc-4ed7e2c85c64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987201] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 3a06a833-5aaa-4b5d-88b3-8a1d469580af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987201] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d2d614f8-15d9-47d6-9e66-8e1f8fcc18da is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.987201] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d8103991-fd1f-4a57-81ce-1a47dc4defe0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987201] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance b4f157a7-350c-4fff-8509-44426714846c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 957.987515] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 374564df-70fa-4b89-8f38-e559245e5ebf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987515] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 7f637326-9388-4d73-a1c7-3e4611ac46c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987581] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987635] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4dcf8335-701d-4386-9de4-f14f5d375d1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987732] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance e0f7c849-315c-4247-a840-d388d48746b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987832] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance ec7f7a46-8dd2-4b75-866f-20e73907f1cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.987933] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.988770] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Migration 8b1f0895-2e84-4d39-9589-aa78a5a0a9b1 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 957.989079] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 2cb53fe5-aa58-479e-9090-0d8509cf164e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.989219] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance a79bbb7e-7b78-4900-9b83-bfce94bb41bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 957.990036] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance b220df77-cd3a-4aeb-87ad-68f56995ec3b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 958.010428] env[65788]: DEBUG nova.network.neutron [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 958.036198] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662745, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.055819] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a270cbe6-16cc-4130-a50e-e8d453048691 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.084445] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ae872d-c26c-4682-99e4-65ca0bceb9d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.100258] env[65788]: INFO nova.compute.manager [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Took 23.17 seconds to build instance. [ 958.116872] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 958.183802] env[65788]: DEBUG nova.network.neutron [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 958.290256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.494534] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 958.535185] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662745, 'name': CloneVM_Task, 'duration_secs': 2.333448} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.535341] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Created linked-clone VM from snapshot [ 958.536148] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769988ae-392b-4de8-a16b-9dd30d596043 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.544872] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Uploading image 8a85504a-ce96-42d0-9709-753767feeca5 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 958.558548] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 958.558873] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7c28d655-52de-4946-aa21-285d47ced996 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.567451] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 958.567451] env[65788]: value = "task-4662749" [ 958.567451] env[65788]: _type = "Task" [ 958.567451] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.576913] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662749, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.602252] env[65788]: DEBUG oslo_concurrency.lockutils [None req-27fabe4c-5499-4c34-950a-93fdff8f2c97 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.680s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.623734] env[65788]: WARNING neutronclient.v2_0.client [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 958.686797] env[65788]: DEBUG oslo_concurrency.lockutils [req-27ffecdb-6dc8-47c6-b30e-c360c14be3b6 req-10c08290-1bc7-43e5-98b3-3aa4deca2658 service nova] Releasing lock "refresh_cache-a79bbb7e-7b78-4900-9b83-bfce94bb41bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 958.687275] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquired lock "refresh_cache-a79bbb7e-7b78-4900-9b83-bfce94bb41bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 958.687468] env[65788]: DEBUG nova.network.neutron [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 958.695749] env[65788]: DEBUG nova.network.neutron [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Port cd2184dd-c826-412e-a259-8e9ebb22c2e5 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 958.956367] env[65788]: DEBUG nova.compute.manager [req-e59cd67b-daf4-48ad-b2ca-0ec5841da6c2 req-d79be5dc-ff90-4de2-aacf-4a20329b8987 service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-vif-deleted-f795ad8a-6471-483d-a6b9-a381bad1c376 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 958.956367] env[65788]: DEBUG nova.compute.manager [req-e59cd67b-daf4-48ad-b2ca-0ec5841da6c2 req-d79be5dc-ff90-4de2-aacf-4a20329b8987 service nova] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Received event network-vif-deleted-0c9e89f7-4194-4cd0-86aa-72b4f8655f1e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 958.956367] env[65788]: DEBUG nova.compute.manager [req-e59cd67b-daf4-48ad-b2ca-0ec5841da6c2 req-d79be5dc-ff90-4de2-aacf-4a20329b8987 service nova] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Received event network-vif-deleted-31daf1d5-0a84-4d17-a694-a7ab16e9b81b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 958.956999] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 958.983270] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 958.983773] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 958.984018] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 958.984258] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 958.984453] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 958.984791] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 958.986206] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.986362] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 958.986589] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 958.987275] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 958.987275] env[65788]: DEBUG nova.virt.hardware [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 958.987925] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9f2cf0-7d45-449b-ab52-43778260907b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.997565] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90e95bc-07a5-4493-9cc3-269a969c7210 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.002854] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 29c0ab17-5df2-4f6c-bdea-3a4a6c857066 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 959.078646] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662749, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.190407] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.190982] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.232412] env[65788]: DEBUG nova.network.neutron [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 959.254143] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.254570] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.344952] env[65788]: WARNING neutronclient.v2_0.client [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 959.345659] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 959.346087] env[65788]: WARNING openstack [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 959.416055] env[65788]: DEBUG nova.network.neutron [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Successfully updated port: f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 959.466212] env[65788]: DEBUG nova.network.neutron [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Updating instance_info_cache with network_info: [{"id": "a0e489a6-9195-4339-a917-3486d9f0da45", "address": "fa:16:3e:92:0d:58", "network": {"id": "9e31867f-356b-4fad-85d2-6cc728aacc64", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1582138937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b693f0eeee664f1ab23b73bb1ea6da0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0e489a6-91", "ovs_interfaceid": "a0e489a6-9195-4339-a917-3486d9f0da45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 959.506872] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance af9120a6-2cea-46da-ba06-6036b9beda78 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 959.507207] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 19 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 959.507370] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4224MB phys_disk=100GB used_disk=19GB total_vcpus=48 used_vcpus=19 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '18', 'num_vm_active': '14', 'num_task_None': '12', 'num_os_type_None': '18', 'num_proj_ffbe44d63c1d432e97849f15615329e1': '2', 'io_workload': '4', 'num_proj_c963c936cf20402d9a65f5e61a11020a': '1', 'num_proj_c4cffe8dfc424459a90e316ac4ac905f': '3', 'num_vm_rescued': '1', 'num_task_deleting': '2', 'num_proj_5e7aa18dcc41403389518d700c2c8b8a': '1', 'num_proj_021bf6dd87e340969a57e7e5b33268e7': '1', 'num_proj_22a3db3230244ef5af1227df6f4a6f12': '1', 'num_proj_2b70d3c8627449eaa6372ebe3bd90233': '1', 'num_task_resize_migrating': '1', 'num_proj_9674e2a5c86b48db8c865a50331ab846': '1', 'num_proj_dd2c665795dc4241a50c46de045f2c80': '1', 'num_proj_94977b75768444138fb914cabddc57c5': '1', 'num_task_image_uploading': '1', 'num_proj_246c0426008f478aa245d006c3c129eb': '1', 'num_proj_7aa1dec53aa4436daa0f3223b92f9b79': '1', 'num_vm_building': '3', 'num_task_spawning': '2', 'num_proj_99b3e41827194188800da4e0544bc19f': '1', 'num_proj_b693f0eeee664f1ab23b73bb1ea6da0a': '1', 'num_proj_3e1713c7002a413fb27180469fded83e': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 959.581320] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662749, 'name': Destroy_Task, 'duration_secs': 0.716361} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.585248] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Destroyed the VM [ 959.585618] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 959.586271] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-76a1f327-49a4-408a-9349-e18195019234 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.593991] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 959.593991] env[65788]: value = "task-4662750" [ 959.593991] env[65788]: _type = "Task" [ 959.593991] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.603559] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662750, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.723925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 959.723925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 959.723925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.883457] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc9c408-66e7-421d-bb87-2cff97f89414 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.893427] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abeae53-371a-423b-9e15-d898613fd187 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.924101] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "refresh_cache-b220df77-cd3a-4aeb-87ad-68f56995ec3b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.924346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "refresh_cache-b220df77-cd3a-4aeb-87ad-68f56995ec3b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 959.924490] env[65788]: DEBUG nova.network.neutron [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 959.926459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b9e8bd-fb40-44be-9b9d-0173135c054f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.935709] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48cc68a7-784b-4c7d-bc73-f3e7db41649d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.951129] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.973096] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Releasing lock "refresh_cache-a79bbb7e-7b78-4900-9b83-bfce94bb41bd" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 959.973096] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Instance network_info: |[{"id": "a0e489a6-9195-4339-a917-3486d9f0da45", "address": "fa:16:3e:92:0d:58", "network": {"id": "9e31867f-356b-4fad-85d2-6cc728aacc64", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1582138937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b693f0eeee664f1ab23b73bb1ea6da0a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0e489a6-91", "ovs_interfaceid": "a0e489a6-9195-4339-a917-3486d9f0da45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 959.973519] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:0d:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6eb7e3e9-5cc2-40f1-a6eb-f70f06531667', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0e489a6-9195-4339-a917-3486d9f0da45', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.981073] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Creating folder: Project (b693f0eeee664f1ab23b73bb1ea6da0a). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 959.981702] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af9c5eb6-e205-40df-8ac1-bfe1afb95290 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.999729] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Created folder: Project (b693f0eeee664f1ab23b73bb1ea6da0a) in parent group-v910111. [ 960.001035] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Creating folder: Instances. Parent ref: group-v910348. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 960.001035] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9ece0ca-5aa7-4792-8e69-dffbdaa194d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.013691] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Created folder: Instances in parent group-v910348. [ 960.014036] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 960.014310] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 960.014682] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a04ea7e-ec9a-4c0d-8219-ff5dd819542f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.039552] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 960.039552] env[65788]: value = "task-4662753" [ 960.039552] env[65788]: _type = "Task" [ 960.039552] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.049830] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662753, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.107492] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662750, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.271956] env[65788]: DEBUG nova.compute.manager [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 960.273036] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8fb375-b4f3-429a-ba16-aaee20f8fa68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.432040] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.432040] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.456309] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 960.469879] env[65788]: DEBUG nova.network.neutron [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 960.489218] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.489637] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.551711] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662753, 'name': CreateVM_Task, 'duration_secs': 0.36225} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.552119] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.552948] env[65788]: WARNING neutronclient.v2_0.client [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.553350] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.553504] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.553869] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 960.554170] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-240fb3b4-46d4-47c2-8608-3853b922b6b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.559942] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 960.559942] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528c17aa-8512-6561-b374-a5fec75dede7" [ 960.559942] env[65788]: _type = "Task" [ 960.559942] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.565565] env[65788]: WARNING neutronclient.v2_0.client [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.566631] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 960.567898] env[65788]: WARNING openstack [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 960.578455] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528c17aa-8512-6561-b374-a5fec75dede7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.605757] env[65788]: DEBUG oslo_vmware.api [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662750, 'name': RemoveSnapshot_Task, 'duration_secs': 0.670192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.606033] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 960.669304] env[65788]: DEBUG nova.network.neutron [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Updating instance_info_cache with network_info: [{"id": "f5aec569-cc68-4c71-a4ca-10710acfbd44", "address": "fa:16:3e:bf:4d:14", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5aec569-cc", "ovs_interfaceid": "f5aec569-cc68-4c71-a4ca-10710acfbd44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 960.730182] env[65788]: WARNING neutronclient.v2_0.client [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 960.764040] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.764040] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 960.764040] env[65788]: DEBUG nova.network.neutron [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 960.786199] env[65788]: INFO nova.compute.manager [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] instance snapshotting [ 960.788964] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d25b378-5796-4f02-adb7-7095acc15658 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.810759] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a58826-1a82-4dd4-aa5e-ba59cd25454f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.962757] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 960.962948] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.037s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.963144] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.441s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.965117] env[65788]: INFO nova.compute.claims [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.999185] env[65788]: DEBUG nova.compute.manager [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Received event network-vif-plugged-f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 960.999290] env[65788]: DEBUG oslo_concurrency.lockutils [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Acquiring lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.000032] env[65788]: DEBUG oslo_concurrency.lockutils [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.000032] env[65788]: DEBUG oslo_concurrency.lockutils [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.000032] env[65788]: DEBUG nova.compute.manager [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] No waiting events found dispatching network-vif-plugged-f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 961.000032] env[65788]: WARNING nova.compute.manager [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Received unexpected event network-vif-plugged-f5aec569-cc68-4c71-a4ca-10710acfbd44 for instance with vm_state building and task_state spawning. [ 961.000229] env[65788]: DEBUG nova.compute.manager [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Received event network-changed-f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 961.000443] env[65788]: DEBUG nova.compute.manager [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Refreshing instance network info cache due to event network-changed-f5aec569-cc68-4c71-a4ca-10710acfbd44. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 961.000616] env[65788]: DEBUG oslo_concurrency.lockutils [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Acquiring lock "refresh_cache-b220df77-cd3a-4aeb-87ad-68f56995ec3b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.072072] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528c17aa-8512-6561-b374-a5fec75dede7, 'name': SearchDatastore_Task, 'duration_secs': 0.011214} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.072460] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 961.072731] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.073022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.073199] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.073420] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.073775] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0136b26e-7455-4b2d-8813-b803d5114ba2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.084488] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.084715] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.085487] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9185b4b5-5f89-4c9d-a046-8e15137e21c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.092264] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 961.092264] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5294d2ae-9a4b-5572-ad52-b38b03fb6fc8" [ 961.092264] env[65788]: _type = "Task" [ 961.092264] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.101782] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5294d2ae-9a4b-5572-ad52-b38b03fb6fc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.110186] env[65788]: WARNING nova.compute.manager [None req-f05c7abc-e744-4653-b765-ed7ddfce6802 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Image not found during snapshot: nova.exception.ImageNotFound: Image 8a85504a-ce96-42d0-9709-753767feeca5 could not be found. [ 961.175839] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "refresh_cache-b220df77-cd3a-4aeb-87ad-68f56995ec3b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 961.175839] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Instance network_info: |[{"id": "f5aec569-cc68-4c71-a4ca-10710acfbd44", "address": "fa:16:3e:bf:4d:14", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5aec569-cc", "ovs_interfaceid": "f5aec569-cc68-4c71-a4ca-10710acfbd44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 961.175839] env[65788]: DEBUG oslo_concurrency.lockutils [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Acquired lock "refresh_cache-b220df77-cd3a-4aeb-87ad-68f56995ec3b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.175839] env[65788]: DEBUG nova.network.neutron [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Refreshing network info cache for port f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 961.175839] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:4d:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20e3f794-c7a3-4696-9488-ecf34c570ef9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5aec569-cc68-4c71-a4ca-10710acfbd44', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.183678] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 961.184923] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.185188] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-baedc1a6-8a28-4558-a0dc-8bbc7deb3ae2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.208269] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.208269] env[65788]: value = "task-4662754" [ 961.208269] env[65788]: _type = "Task" [ 961.208269] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.217637] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662754, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.266531] env[65788]: WARNING neutronclient.v2_0.client [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.267305] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.267654] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.322949] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 961.324042] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-087da01f-03c3-4ac8-98d8-9ba400ed586d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.334472] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 961.334472] env[65788]: value = "task-4662755" [ 961.334472] env[65788]: _type = "Task" [ 961.334472] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.345552] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662755, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.399856] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.400361] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.473072] env[65788]: WARNING neutronclient.v2_0.client [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.473875] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.474276] env[65788]: WARNING openstack [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.485296] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "e0f7c849-315c-4247-a840-d388d48746b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.485296] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "e0f7c849-315c-4247-a840-d388d48746b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.486244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "e0f7c849-315c-4247-a840-d388d48746b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.486244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "e0f7c849-315c-4247-a840-d388d48746b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.486528] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "e0f7c849-315c-4247-a840-d388d48746b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.488733] env[65788]: INFO nova.compute.manager [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Terminating instance [ 961.577982] env[65788]: DEBUG nova.network.neutron [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance_info_cache with network_info: [{"id": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "address": "fa:16:3e:0a:42:08", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2184dd-c8", "ovs_interfaceid": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 961.605983] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5294d2ae-9a4b-5572-ad52-b38b03fb6fc8, 'name': SearchDatastore_Task, 'duration_secs': 0.011609} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.606329] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bbe318e-e4f0-445c-844f-ba0df14d9b24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.613538] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 961.613538] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a516c3-a77f-ffad-4d4f-1f08509026fe" [ 961.613538] env[65788]: _type = "Task" [ 961.613538] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.624898] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a516c3-a77f-ffad-4d4f-1f08509026fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.686183] env[65788]: WARNING neutronclient.v2_0.client [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.687079] env[65788]: WARNING openstack [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.687606] env[65788]: WARNING openstack [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.719926] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662754, 'name': CreateVM_Task, 'duration_secs': 0.421042} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.719926] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 961.720950] env[65788]: WARNING neutronclient.v2_0.client [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.720950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.720950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 961.721208] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 961.721480] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abdc8fc8-8735-4660-918d-dedeb5879a76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.727639] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 961.727639] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520be373-1e82-3d4e-6d12-30060210420d" [ 961.727639] env[65788]: _type = "Task" [ 961.727639] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.740017] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520be373-1e82-3d4e-6d12-30060210420d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.799028] env[65788]: WARNING openstack [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.799289] env[65788]: WARNING openstack [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.845786] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662755, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.868289] env[65788]: WARNING neutronclient.v2_0.client [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 961.869186] env[65788]: WARNING openstack [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 961.869562] env[65788]: WARNING openstack [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 961.970629] env[65788]: DEBUG nova.network.neutron [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Updated VIF entry in instance network info cache for port f5aec569-cc68-4c71-a4ca-10710acfbd44. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 961.971030] env[65788]: DEBUG nova.network.neutron [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Updating instance_info_cache with network_info: [{"id": "f5aec569-cc68-4c71-a4ca-10710acfbd44", "address": "fa:16:3e:bf:4d:14", "network": {"id": "c61b5c79-c2fb-488b-9a62-64d9cbde9891", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1424731087-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3e1713c7002a413fb27180469fded83e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20e3f794-c7a3-4696-9488-ecf34c570ef9", "external-id": "nsx-vlan-transportzone-509", "segmentation_id": 509, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5aec569-cc", "ovs_interfaceid": "f5aec569-cc68-4c71-a4ca-10710acfbd44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 961.995206] env[65788]: DEBUG nova.compute.manager [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 961.995868] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.996820] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aba8ca-190f-44fb-83ab-8a27fa625864 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.006469] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.006765] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d62fd91-eefb-4729-968c-c5e206dde70e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.014915] env[65788]: DEBUG oslo_vmware.api [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 962.014915] env[65788]: value = "task-4662756" [ 962.014915] env[65788]: _type = "Task" [ 962.014915] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.029336] env[65788]: DEBUG oslo_vmware.api [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662756, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.081199] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.133888] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a516c3-a77f-ffad-4d4f-1f08509026fe, 'name': SearchDatastore_Task, 'duration_secs': 0.010526} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.134588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.134822] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a79bbb7e-7b78-4900-9b83-bfce94bb41bd/a79bbb7e-7b78-4900-9b83-bfce94bb41bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.135256] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8dbb075-db28-4826-8135-39d548ee0bcd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.149825] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 962.149825] env[65788]: value = "task-4662757" [ 962.149825] env[65788]: _type = "Task" [ 962.149825] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.168972] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662757, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.247058] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520be373-1e82-3d4e-6d12-30060210420d, 'name': SearchDatastore_Task, 'duration_secs': 0.012971} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.247497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.247787] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.248108] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.248563] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 962.248563] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.248774] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-590af22f-f94a-4ba0-8b19-006a046bbe93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.271209] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.271424] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.272314] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c23788b4-da31-43df-88cb-5b3f1c42314c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.280121] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 962.280121] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52af0f4a-d0f9-e194-65cd-0dec5f9dc53d" [ 962.280121] env[65788]: _type = "Task" [ 962.280121] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.292894] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52af0f4a-d0f9-e194-65cd-0dec5f9dc53d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.346455] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662755, 'name': CreateSnapshot_Task, 'duration_secs': 0.909653} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.346455] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 962.347275] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7dd897-d2b8-489c-8452-b301ca04ada8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.401673] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8d257d-0ec1-472a-bfd5-140411cbccf0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.413918] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b672e1-4de1-4c3a-bc4e-b33bbd7278fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.452494] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aa4342-3e28-4241-9b63-4b759ec1219f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.464887] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3372b439-9ad1-4374-8332-3c11d51e067e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.475050] env[65788]: DEBUG oslo_concurrency.lockutils [req-586f5e99-8796-4044-b9fa-1546b278a826 req-2416c08c-5164-43d4-8ab2-01e3f4c3a4e9 service nova] Releasing lock "refresh_cache-b220df77-cd3a-4aeb-87ad-68f56995ec3b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.485061] env[65788]: DEBUG nova.compute.provider_tree [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.530552] env[65788]: DEBUG oslo_vmware.api [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662756, 'name': PowerOffVM_Task, 'duration_secs': 0.198519} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.530899] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.531091] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 962.531429] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97cec20f-8f04-4ef1-ad62-a46341fc5850 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.609228] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b42d8b-318f-4eae-8976-7f25bab5a3ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.633732] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7bf8ac-00d4-44b2-8676-bcb11ab0c5a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.636808] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.637042] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.637218] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleting the datastore file [datastore1] e0f7c849-315c-4247-a840-d388d48746b9 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.637497] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8ea1777-0af5-4d0c-a826-36be5796717f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.647062] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.653455] env[65788]: DEBUG oslo_vmware.api [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 962.653455] env[65788]: value = "task-4662759" [ 962.653455] env[65788]: _type = "Task" [ 962.653455] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.669025] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662757, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.672881] env[65788]: DEBUG oslo_vmware.api [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.793990] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52af0f4a-d0f9-e194-65cd-0dec5f9dc53d, 'name': SearchDatastore_Task, 'duration_secs': 0.014944} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.794854] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0764d8f8-ce1f-4989-be81-40be29a8106d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.803245] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 962.803245] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a2dbdb-1947-2b58-1022-e458a00c644e" [ 962.803245] env[65788]: _type = "Task" [ 962.803245] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.814444] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a2dbdb-1947-2b58-1022-e458a00c644e, 'name': SearchDatastore_Task, 'duration_secs': 0.010207} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.814671] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 962.814931] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b220df77-cd3a-4aeb-87ad-68f56995ec3b/b220df77-cd3a-4aeb-87ad-68f56995ec3b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.815263] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db021562-c691-485f-a931-7e9d29b2f7eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.824847] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 962.824847] env[65788]: value = "task-4662760" [ 962.824847] env[65788]: _type = "Task" [ 962.824847] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.833425] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.872381] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 962.872827] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ce7fb80d-9d13-4837-80ea-1220ad58b077 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.882602] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 962.882602] env[65788]: value = "task-4662761" [ 962.882602] env[65788]: _type = "Task" [ 962.882602] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.894546] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.988183] env[65788]: DEBUG nova.scheduler.client.report [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 963.156516] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.160937] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d8edfe0-3a81-4bb2-ac42-284ed615dccb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.175585] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662757, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.596354} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.182324] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] a79bbb7e-7b78-4900-9b83-bfce94bb41bd/a79bbb7e-7b78-4900-9b83-bfce94bb41bd.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.182774] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.184310] env[65788]: DEBUG oslo_vmware.api [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.184683] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 963.184683] env[65788]: value = "task-4662762" [ 963.184683] env[65788]: _type = "Task" [ 963.184683] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.184940] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df55e83a-0dd2-4c33-a79c-8e1492aa2d53 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.187542] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.187814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.188086] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.188341] env[65788]: INFO nova.compute.manager [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Took 1.19 seconds to destroy the instance on the hypervisor. [ 963.188702] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 963.189098] env[65788]: DEBUG nova.compute.manager [-] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 963.189209] env[65788]: DEBUG nova.network.neutron [-] [instance: e0f7c849-315c-4247-a840-d388d48746b9] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 963.189565] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.190309] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 963.190779] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 963.210439] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 963.210439] env[65788]: value = "task-4662763" [ 963.210439] env[65788]: _type = "Task" [ 963.210439] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.214872] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662762, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.229580] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.247257] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 963.335025] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662760, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508271} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.336292] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b220df77-cd3a-4aeb-87ad-68f56995ec3b/b220df77-cd3a-4aeb-87ad-68f56995ec3b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.336292] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.336292] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-630f4af9-c34e-49e7-88f7-83c8ce6a9f63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.345940] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 963.345940] env[65788]: value = "task-4662764" [ 963.345940] env[65788]: _type = "Task" [ 963.345940] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.357253] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.395688] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.497536] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.497536] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 963.509177] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.387s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.513088] env[65788]: INFO nova.compute.claims [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.592270] env[65788]: DEBUG nova.compute.manager [req-db334251-cd09-40eb-9067-1f849b0069f9 req-df4aaa92-22d5-45ee-8b3b-87f2e0af19ee service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Received event network-vif-deleted-00af96db-551c-48f3-9b82-c272038d58a8 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 963.592475] env[65788]: INFO nova.compute.manager [req-db334251-cd09-40eb-9067-1f849b0069f9 req-df4aaa92-22d5-45ee-8b3b-87f2e0af19ee service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Neutron deleted interface 00af96db-551c-48f3-9b82-c272038d58a8; detaching it from the instance and deleting it from the info cache [ 963.592701] env[65788]: DEBUG nova.network.neutron [req-db334251-cd09-40eb-9067-1f849b0069f9 req-df4aaa92-22d5-45ee-8b3b-87f2e0af19ee service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 963.700208] env[65788]: DEBUG oslo_vmware.api [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662762, 'name': PowerOnVM_Task, 'duration_secs': 0.5201} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.700756] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.701017] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66236279-b07f-465f-88dd-34ad738900b1 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance '2cb53fe5-aa58-479e-9090-0d8509cf164e' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 963.724824] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.170232} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.725558] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.725926] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2571dc6-ae99-493b-a61e-3fa0405546da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.750231] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] a79bbb7e-7b78-4900-9b83-bfce94bb41bd/a79bbb7e-7b78-4900-9b83-bfce94bb41bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.751395] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5979d2e4-57de-4d44-92a2-1c2e1afa3f4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.772137] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 963.772137] env[65788]: value = "task-4662765" [ 963.772137] env[65788]: _type = "Task" [ 963.772137] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.781986] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.857675] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078533} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.858012] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.858904] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605eb07b-74c2-4efd-a456-5a08dbf43584 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.883267] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] b220df77-cd3a-4aeb-87ad-68f56995ec3b/b220df77-cd3a-4aeb-87ad-68f56995ec3b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.883454] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-503c9a9d-7e91-4232-9470-894f7195f757 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.908719] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.910362] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 963.910362] env[65788]: value = "task-4662766" [ 963.910362] env[65788]: _type = "Task" [ 963.910362] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.920496] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662766, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.017190] env[65788]: DEBUG nova.network.neutron [-] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 964.020095] env[65788]: DEBUG nova.compute.utils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 964.021971] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.516s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 964.022868] env[65788]: DEBUG nova.compute.utils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Instance ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c could not be found. {{(pid=65788) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 964.024782] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 964.024782] env[65788]: DEBUG nova.network.neutron [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 964.024782] env[65788]: WARNING neutronclient.v2_0.client [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 964.025048] env[65788]: WARNING neutronclient.v2_0.client [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 964.026606] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.026606] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.035618] env[65788]: DEBUG nova.compute.manager [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Instance disappeared during build. {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2557}} [ 964.036416] env[65788]: DEBUG nova.compute.manager [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Unplugging VIFs for instance {{(pid=65788) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3051}} [ 964.036416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.036416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.037115] env[65788]: DEBUG nova.network.neutron [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 964.038695] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.841s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.040349] env[65788]: INFO nova.compute.claims [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.096046] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76c85595-ae7d-4921-bf06-3bc4c4706a30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.111452] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08949c61-6e47-4805-8829-395388b0b468 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.126269] env[65788]: DEBUG nova.policy [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d53d46981ae41fe88f5c039b1a09854', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a395530cf6f643d6a727bb2f44a77491', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 964.156196] env[65788]: DEBUG nova.compute.manager [req-db334251-cd09-40eb-9067-1f849b0069f9 req-df4aaa92-22d5-45ee-8b3b-87f2e0af19ee service nova] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Detach interface failed, port_id=00af96db-551c-48f3-9b82-c272038d58a8, reason: Instance e0f7c849-315c-4247-a840-d388d48746b9 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 964.283728] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662765, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.409030] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.424524] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662766, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.434466] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.434710] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.481662] env[65788]: DEBUG nova.network.neutron [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Successfully created port: ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 964.523084] env[65788]: INFO nova.compute.manager [-] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Took 1.33 seconds to deallocate network for instance. [ 964.539397] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 964.545099] env[65788]: DEBUG nova.compute.utils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Can not refresh info_cache because instance was not found {{(pid=65788) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 964.545099] env[65788]: WARNING neutronclient.v2_0.client [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: nova.exception_Remote.InstanceNotFound_Remote: Instance ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c could not be found. [ 964.547302] env[65788]: WARNING openstack [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 964.547776] env[65788]: WARNING openstack [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 964.585019] env[65788]: DEBUG nova.network.neutron [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 964.681259] env[65788]: DEBUG nova.network.neutron [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 964.734688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.734688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.734688] env[65788]: INFO nova.compute.manager [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Shelving [ 964.784475] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662765, 'name': ReconfigVM_Task, 'duration_secs': 0.838142} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.784786] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Reconfigured VM instance instance-0000004f to attach disk [datastore2] a79bbb7e-7b78-4900-9b83-bfce94bb41bd/a79bbb7e-7b78-4900-9b83-bfce94bb41bd.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.785582] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ac2d9f2-ea47-4323-bb41-945d3aded34b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.794459] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 964.794459] env[65788]: value = "task-4662767" [ 964.794459] env[65788]: _type = "Task" [ 964.794459] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.807247] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662767, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.910559] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.922614] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662766, 'name': ReconfigVM_Task, 'duration_secs': 0.724312} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.922959] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Reconfigured VM instance instance-00000050 to attach disk [datastore2] b220df77-cd3a-4aeb-87ad-68f56995ec3b/b220df77-cd3a-4aeb-87ad-68f56995ec3b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.923665] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e495d79-ea18-4923-99fb-a0fe9f33418a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.932293] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 964.932293] env[65788]: value = "task-4662768" [ 964.932293] env[65788]: _type = "Task" [ 964.932293] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.938415] env[65788]: DEBUG nova.compute.utils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 964.943857] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662768, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.039359] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.184578] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 965.184643] env[65788]: DEBUG nova.compute.manager [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=65788) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3074}} [ 965.184820] env[65788]: DEBUG nova.compute.manager [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 965.185100] env[65788]: DEBUG nova.network.neutron [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 965.185576] env[65788]: WARNING neutronclient.v2_0.client [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: nova.exception_Remote.InstanceNotFound_Remote: Instance ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c could not be found. [ 965.186616] env[65788]: WARNING openstack [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 965.186916] env[65788]: WARNING openstack [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 965.212277] env[65788]: DEBUG nova.network.neutron [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 965.212768] env[65788]: WARNING neutronclient.v2_0.client [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: nova.exception_Remote.InstanceNotFound_Remote: Instance ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c could not be found. [ 965.308349] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662767, 'name': Rename_Task, 'duration_secs': 0.228854} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.311383] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.311879] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2d163c0-a510-4ad1-82fa-547dc6f3f5c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.319911] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 965.319911] env[65788]: value = "task-4662769" [ 965.319911] env[65788]: _type = "Task" [ 965.319911] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.332589] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662769, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.412113] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.444585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.445246] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662768, 'name': Rename_Task, 'duration_secs': 0.191816} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.445567] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.446530] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b461028-7264-4ed8-9057-dd2b58483819 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.450351] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f5dbe4-5eea-441b-bda8-4c1b61e29f3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.462642] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58108893-e121-4701-9229-38052d86e69f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.466496] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 965.466496] env[65788]: value = "task-4662770" [ 965.466496] env[65788]: _type = "Task" [ 965.466496] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.502563] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2247836d-7130-46c8-8eac-ee619eea15f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.509254] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662770, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.516844] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c300270-5545-4951-905f-789bce316aae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.534919] env[65788]: DEBUG nova.compute.provider_tree [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.572636] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 965.603243] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 965.603605] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 965.603818] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 965.604038] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 965.604201] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 965.604365] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 965.604581] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.604739] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 965.604922] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 965.605146] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 965.605443] env[65788]: DEBUG nova.virt.hardware [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 965.606419] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7392a3d-c336-49fb-b8df-1432a6a9a7fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.616770] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a23e3e8-78fc-4177-8b4b-6f1755677732 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.715625] env[65788]: DEBUG nova.network.neutron [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 965.743036] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.743425] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-910cbaa0-af2f-4843-99bb-e320e0a38331 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.753058] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 965.753058] env[65788]: value = "task-4662771" [ 965.753058] env[65788]: _type = "Task" [ 965.753058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.763303] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.836544] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662769, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.913215] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662761, 'name': CloneVM_Task, 'duration_secs': 2.670171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.913576] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Created linked-clone VM from snapshot [ 965.914698] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e56d1c-4af1-4555-af31-945e65a8eadd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.925539] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Uploading image 0b0e4bc3-60eb-4ac5-b652-2cc27f204568 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 965.958175] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 965.958175] env[65788]: value = "vm-910353" [ 965.958175] env[65788]: _type = "VirtualMachine" [ 965.958175] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 965.958546] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2f2ceb2b-7b07-4a58-b3ac-6bbfdac153db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.969306] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lease: (returnval){ [ 965.969306] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e890d-ee3f-64a5-08e3-51209c53c86d" [ 965.969306] env[65788]: _type = "HttpNfcLease" [ 965.969306] env[65788]: } obtained for exporting VM: (result){ [ 965.969306] env[65788]: value = "vm-910353" [ 965.969306] env[65788]: _type = "VirtualMachine" [ 965.969306] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 965.969856] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the lease: (returnval){ [ 965.969856] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e890d-ee3f-64a5-08e3-51209c53c86d" [ 965.969856] env[65788]: _type = "HttpNfcLease" [ 965.969856] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 965.982770] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 965.982770] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e890d-ee3f-64a5-08e3-51209c53c86d" [ 965.982770] env[65788]: _type = "HttpNfcLease" [ 965.982770] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 965.986677] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662770, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.039260] env[65788]: DEBUG nova.scheduler.client.report [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 966.202029] env[65788]: DEBUG nova.network.neutron [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Successfully updated port: ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 966.218938] env[65788]: INFO nova.compute.manager [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c] Took 1.03 seconds to deallocate network for instance. [ 966.263986] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662771, 'name': PowerOffVM_Task, 'duration_secs': 0.224212} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.264414] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.265401] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6bb840-b4f3-4626-bfb4-f37606a511f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.270530] env[65788]: DEBUG nova.compute.manager [req-89818af5-7f0d-4989-8961-a61d4c6fc749 req-dca82cae-3ebb-4ab4-9b85-162c6e01cd5a service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Received event network-vif-plugged-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 966.270858] env[65788]: DEBUG oslo_concurrency.lockutils [req-89818af5-7f0d-4989-8961-a61d4c6fc749 req-dca82cae-3ebb-4ab4-9b85-162c6e01cd5a service nova] Acquiring lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.271132] env[65788]: DEBUG oslo_concurrency.lockutils [req-89818af5-7f0d-4989-8961-a61d4c6fc749 req-dca82cae-3ebb-4ab4-9b85-162c6e01cd5a service nova] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.271353] env[65788]: DEBUG oslo_concurrency.lockutils [req-89818af5-7f0d-4989-8961-a61d4c6fc749 req-dca82cae-3ebb-4ab4-9b85-162c6e01cd5a service nova] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.271517] env[65788]: DEBUG nova.compute.manager [req-89818af5-7f0d-4989-8961-a61d4c6fc749 req-dca82cae-3ebb-4ab4-9b85-162c6e01cd5a service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] No waiting events found dispatching network-vif-plugged-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 966.271687] env[65788]: WARNING nova.compute.manager [req-89818af5-7f0d-4989-8961-a61d4c6fc749 req-dca82cae-3ebb-4ab4-9b85-162c6e01cd5a service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Received unexpected event network-vif-plugged-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b for instance with vm_state building and task_state spawning. [ 966.287961] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbad4113-4f5c-4827-86b4-1af3a6d70f1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.333789] env[65788]: DEBUG oslo_vmware.api [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662769, 'name': PowerOnVM_Task, 'duration_secs': 0.721947} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.334084] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.334308] env[65788]: INFO nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Took 9.95 seconds to spawn the instance on the hypervisor. [ 966.334496] env[65788]: DEBUG nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 966.335407] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d346b70-9ddf-4a9e-97fe-39306ea8f1fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.481561] env[65788]: DEBUG oslo_vmware.api [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662770, 'name': PowerOnVM_Task, 'duration_secs': 0.586245} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.482150] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.482433] env[65788]: INFO nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Took 7.53 seconds to spawn the instance on the hypervisor. [ 966.482549] env[65788]: DEBUG nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 966.483423] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa5f89b-0df3-45e1-99bd-8dee956ee56b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.488803] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 966.488803] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e890d-ee3f-64a5-08e3-51209c53c86d" [ 966.488803] env[65788]: _type = "HttpNfcLease" [ 966.488803] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 966.489235] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 966.489235] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529e890d-ee3f-64a5-08e3-51209c53c86d" [ 966.489235] env[65788]: _type = "HttpNfcLease" [ 966.489235] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 966.490052] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cea6dba-4f50-414e-a72a-8cd11bbc6d79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.504869] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5274a35e-6215-d418-e069-b7938364d500/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 966.505150] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5274a35e-6215-d418-e069-b7938364d500/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 966.563532] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.564020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.565714] env[65788]: DEBUG nova.compute.manager [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Going to confirm migration 5 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 966.569021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.569021] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 966.570113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.570113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.570402] env[65788]: INFO nova.compute.manager [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Attaching volume 86b8defd-6720-439f-9a6c-b759609a2da3 to /dev/sdb [ 966.578062] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.602s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.578062] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.578254] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.467s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.578547] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.581473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.738s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.581825] env[65788]: INFO nova.compute.claims [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.613390] env[65788]: INFO nova.scheduler.client.report [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Deleted allocations for instance d2d614f8-15d9-47d6-9e66-8e1f8fcc18da [ 966.615277] env[65788]: INFO nova.scheduler.client.report [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Deleted allocations for instance b4f157a7-350c-4fff-8509-44426714846c [ 966.625890] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-373973fc-30e8-4ff5-bb03-4104ad5c5cc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.630089] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c728dd0-562e-4eb6-940d-06c0b1a8a234 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.639657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3491938-fc77-4428-912e-c72c36b5d737 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.656747] env[65788]: DEBUG nova.virt.block_device [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating existing volume attachment record: 9fcec797-e493-4562-aa9c-65f5209801be {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 966.704328] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.704586] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquired lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 966.704858] env[65788]: DEBUG nova.network.neutron [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 966.803240] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 966.803848] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0334f518-824a-450c-9ea3-d20151e949aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.817882] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 966.817882] env[65788]: value = "task-4662774" [ 966.817882] env[65788]: _type = "Task" [ 966.817882] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.829047] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662774, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.854380] env[65788]: INFO nova.compute.manager [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Took 27.57 seconds to build instance. [ 967.009956] env[65788]: INFO nova.compute.manager [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Took 25.53 seconds to build instance. [ 967.078137] env[65788]: DEBUG nova.compute.utils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 967.080619] env[65788]: WARNING neutronclient.v2_0.client [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.083759] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 967.084307] env[65788]: DEBUG nova.network.neutron [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 967.085673] env[65788]: WARNING neutronclient.v2_0.client [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.087047] env[65788]: WARNING neutronclient.v2_0.client [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.087293] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.088010] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.099340] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.124458] env[65788]: DEBUG oslo_concurrency.lockutils [None req-241b2b70-e0ab-4743-882c-bd91a8f4373b tempest-ListServerFiltersTestJSON-326822078 tempest-ListServerFiltersTestJSON-326822078-project-member] Lock "d2d614f8-15d9-47d6-9e66-8e1f8fcc18da" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.739s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.138661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f26a3d0-97c4-47c9-b05b-c924ceb0b163 tempest-ServerRescueTestJSONUnderV235-739854739 tempest-ServerRescueTestJSONUnderV235-739854739-project-member] Lock "b4f157a7-350c-4fff-8509-44426714846c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.753s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.178975] env[65788]: DEBUG nova.policy [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32b4adcc8fa42fba0dbffd4582fcf69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36859c1eb994614b2a77400f811cf50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 967.208806] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.209784] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.225864] env[65788]: WARNING neutronclient.v2_0.client [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 967.226343] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.226656] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 967.226897] env[65788]: DEBUG nova.network.neutron [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 967.227165] env[65788]: DEBUG nova.objects.instance [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'info_cache' on Instance uuid 2cb53fe5-aa58-479e-9090-0d8509cf164e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.250046] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4bb43e1d-208e-4e09-9e80-130f8934b2f9 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "ce49ac6b-6be2-47d7-8d9c-b308f95e0d3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.162s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.334086] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662774, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.358079] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b50d833-847f-442e-9fbf-24ebd083da23 tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.080s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.358534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.260s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.358877] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.359233] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.359915] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.362254] env[65788]: INFO nova.compute.manager [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Terminating instance [ 967.428413] env[65788]: DEBUG nova.network.neutron [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 967.513081] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1a7de155-9e6c-47d8-a3af-d687f969b409 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.039s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.546735] env[65788]: DEBUG nova.network.neutron [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Successfully created port: 9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 967.597685] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 967.834201] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662774, 'name': CreateSnapshot_Task, 'duration_secs': 0.773187} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.836017] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 967.836017] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0def4885-ca2d-4f45-9a79-0ed210ad1f0b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.870172] env[65788]: DEBUG nova.compute.manager [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 967.870481] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.877512] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d790fa5-cb36-474b-9a93-7a32b34066af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.887684] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.888055] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84170427-1ddb-4c6e-bfc0-32edca42dc67 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.896390] env[65788]: DEBUG oslo_vmware.api [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 967.896390] env[65788]: value = "task-4662777" [ 967.896390] env[65788]: _type = "Task" [ 967.896390] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.901512] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 967.901931] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 967.924650] env[65788]: DEBUG oslo_vmware.api [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.009060] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc73ea42-c104-43e5-bcd3-4e2632fc88d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.018670] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb696da-4ee8-409c-b872-e744bcd6eff5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.058091] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9996ab59-3290-4602-9037-9ac7fb265e14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.072027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c32714-bf68-4832-8e2d-4be251ac2dad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.095350] env[65788]: DEBUG nova.compute.provider_tree [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.243829] env[65788]: WARNING neutronclient.v2_0.client [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 968.244966] env[65788]: WARNING openstack [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 968.245568] env[65788]: WARNING openstack [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 968.255717] env[65788]: DEBUG nova.compute.manager [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 968.257702] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63dc7db-fb2b-49c7-9cbb-51bb8cf62e78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.359211] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 968.359622] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f2a557b1-4d35-4d0c-97ba-cac8197883ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.371781] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 968.371781] env[65788]: value = "task-4662778" [ 968.371781] env[65788]: _type = "Task" [ 968.371781] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.385943] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662778, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.409070] env[65788]: DEBUG oslo_vmware.api [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662777, 'name': PowerOffVM_Task, 'duration_secs': 0.227101} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.409440] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.409630] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.409932] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4d296d1-94c7-4157-8ca6-a85d4e567406 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.485901] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.486244] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.486447] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Deleting the datastore file [datastore2] a79bbb7e-7b78-4900-9b83-bfce94bb41bd {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.486769] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d2270ed-7d6c-493c-93f9-55bc8368d41c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.497089] env[65788]: DEBUG oslo_vmware.api [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for the task: (returnval){ [ 968.497089] env[65788]: value = "task-4662780" [ 968.497089] env[65788]: _type = "Task" [ 968.497089] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.506752] env[65788]: DEBUG oslo_vmware.api [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.599979] env[65788]: DEBUG nova.scheduler.client.report [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 968.613356] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 968.643131] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 968.643423] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 968.643604] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 968.643912] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 968.644204] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 968.644376] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 968.644589] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.644892] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 968.645235] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 968.645329] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 968.645482] env[65788]: DEBUG nova.virt.hardware [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 968.646405] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e02bde6-9bae-4db2-bdbe-71d9a8492947 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.656961] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c85519-23ce-4cd5-a021-cabc1243adb7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.775381] env[65788]: INFO nova.compute.manager [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] instance snapshotting [ 968.778882] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2833bc-4248-43ff-95e5-c83cec633c35 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.799832] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7c533d-d061-49c9-94a6-3e274e17e30b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.883098] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662778, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.902066] env[65788]: WARNING neutronclient.v2_0.client [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 968.902813] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 968.903210] env[65788]: WARNING openstack [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.011222] env[65788]: DEBUG oslo_vmware.api [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Task: {'id': task-4662780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233634} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.011222] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.011568] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.012039] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.012039] env[65788]: INFO nova.compute.manager [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Took 1.14 seconds to destroy the instance on the hypervisor. [ 969.012457] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 969.012583] env[65788]: DEBUG nova.compute.manager [-] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 969.012768] env[65788]: DEBUG nova.network.neutron [-] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 969.013025] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.013709] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.013960] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.106503] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.107116] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 969.110639] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.320s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.110958] env[65788]: DEBUG nova.objects.instance [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lazy-loading 'resources' on Instance uuid 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.314708] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 969.314708] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b1e95868-0a1d-4501-9447-150203c38a28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.324072] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 969.324072] env[65788]: value = "task-4662782" [ 969.324072] env[65788]: _type = "Task" [ 969.324072] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.335666] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662782, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.380768] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.386698] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662778, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.397884] env[65788]: WARNING openstack [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.398407] env[65788]: WARNING openstack [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.581980] env[65788]: DEBUG nova.network.neutron [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Successfully updated port: 9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 969.619111] env[65788]: DEBUG nova.compute.utils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 969.624247] env[65788]: DEBUG nova.network.neutron [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Updating instance_info_cache with network_info: [{"id": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "address": "fa:16:3e:6c:31:d3", "network": {"id": "6e4a5321-3d4d-4146-ad63-556e8907c3a7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1463785200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a395530cf6f643d6a727bb2f44a77491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec66b0ca-3c", "ovs_interfaceid": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 969.626129] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 969.626521] env[65788]: DEBUG nova.network.neutron [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 969.626906] env[65788]: WARNING neutronclient.v2_0.client [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.627649] env[65788]: WARNING neutronclient.v2_0.client [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.628045] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.632048] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.637334] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 969.644703] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Releasing lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 969.644703] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Instance network_info: |[{"id": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "address": "fa:16:3e:6c:31:d3", "network": {"id": "6e4a5321-3d4d-4146-ad63-556e8907c3a7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1463785200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a395530cf6f643d6a727bb2f44a77491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec66b0ca-3c", "ovs_interfaceid": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 969.644703] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:31:d3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb68953b-dee5-4d9d-b47b-277336ba76dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.654699] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Creating folder: Project (a395530cf6f643d6a727bb2f44a77491). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.656279] env[65788]: DEBUG nova.scheduler.client.report [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 969.664217] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7f9ad7ef-e4e8-4d27-88d9-9bd2e74560cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.667442] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.667773] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.668611] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.668898] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.669154] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.671336] env[65788]: INFO nova.compute.manager [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Terminating instance [ 969.680240] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Created folder: Project (a395530cf6f643d6a727bb2f44a77491) in parent group-v910111. [ 969.680520] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Creating folder: Instances. Parent ref: group-v910358. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.680836] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-755dfbf7-35f1-42f6-9000-33664cb37dc2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.688959] env[65788]: DEBUG nova.scheduler.client.report [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 969.690420] env[65788]: DEBUG nova.compute.provider_tree [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 969.693926] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Created folder: Instances in parent group-v910358. [ 969.694188] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 969.694520] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.694729] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-491ccfde-2a46-4fe4-b2e4-e39fb84f27d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.715729] env[65788]: DEBUG nova.scheduler.client.report [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 969.725045] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.725045] env[65788]: value = "task-4662785" [ 969.725045] env[65788]: _type = "Task" [ 969.725045] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.740150] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662785, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.742626] env[65788]: DEBUG nova.scheduler.client.report [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 969.794552] env[65788]: DEBUG nova.policy [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '105258eea35d4e22aa4914ec5b5ccbdf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22a3db3230244ef5af1227df6f4a6f12', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 969.839156] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662782, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.841138] env[65788]: WARNING neutronclient.v2_0.client [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 969.842110] env[65788]: WARNING openstack [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 969.842712] env[65788]: WARNING openstack [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 969.885674] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662778, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.001366] env[65788]: DEBUG nova.compute.manager [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Received event network-changed-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 970.001691] env[65788]: DEBUG nova.compute.manager [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Refreshing instance network info cache due to event network-changed-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 970.001817] env[65788]: DEBUG oslo_concurrency.lockutils [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Acquiring lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.002262] env[65788]: DEBUG oslo_concurrency.lockutils [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Acquired lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.002262] env[65788]: DEBUG nova.network.neutron [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Refreshing network info cache for port ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 970.083899] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.084534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.084756] env[65788]: DEBUG nova.network.neutron [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 970.097412] env[65788]: DEBUG nova.network.neutron [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance_info_cache with network_info: [{"id": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "address": "fa:16:3e:0a:42:08", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2184dd-c8", "ovs_interfaceid": "cd2184dd-c826-412e-a259-8e9ebb22c2e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.176205] env[65788]: DEBUG nova.compute.manager [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 970.176498] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.180019] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618a68c7-5959-4d00-9921-73f78518b23e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.202207] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.202207] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c43979c-5452-4f09-b271-bbbdeb7ec4cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.218877] env[65788]: DEBUG oslo_vmware.api [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 970.218877] env[65788]: value = "task-4662786" [ 970.218877] env[65788]: _type = "Task" [ 970.218877] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.227040] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fc2734-4329-4a4e-88bf-73cb840d75cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.238429] env[65788]: DEBUG oslo_vmware.api [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662786, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.252800] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f30071-e6c1-4326-8796-e50567db9507 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.257168] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662785, 'name': CreateVM_Task, 'duration_secs': 0.472803} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.257168] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.257886] env[65788]: WARNING neutronclient.v2_0.client [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.258299] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.258462] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.258818] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 970.260838] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a34a0f62-c076-4451-a111-3dac82cf1a0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.290766] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c021b803-4978-433e-92f7-533d86f6cd63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.296919] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 970.296919] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d5a679-f90f-170b-d489-082148446a97" [ 970.296919] env[65788]: _type = "Task" [ 970.296919] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.305278] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76147f44-fad0-49f6-8d82-511400cc830f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.314543] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d5a679-f90f-170b-d489-082148446a97, 'name': SearchDatastore_Task, 'duration_secs': 0.01429} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.315304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.315463] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.315761] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.316014] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.316197] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.316571] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95ff1253-67f6-4588-b757-ea0591ccea79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.330399] env[65788]: DEBUG nova.compute.provider_tree [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.335962] env[65788]: DEBUG nova.network.neutron [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Successfully created port: 5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 970.343557] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.343557] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.344447] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662782, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.345553] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75d74d67-24cf-49c4-8be1-45a7a7c4eff2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.353352] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 970.353352] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218a9de-db80-f5d4-70d8-aef52be820b7" [ 970.353352] env[65788]: _type = "Task" [ 970.353352] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.363303] env[65788]: DEBUG nova.network.neutron [-] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 970.364940] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218a9de-db80-f5d4-70d8-aef52be820b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.388178] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662778, 'name': CloneVM_Task, 'duration_secs': 1.640004} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.388178] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Created linked-clone VM from snapshot [ 970.388178] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98f79fe-b7d4-4977-ae40-58cbf07bec30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.398352] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Uploading image 0f208f1a-1c47-4d89-ac7d-a4f584ec31ed {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 970.419381] env[65788]: DEBUG nova.compute.manager [req-1e0d607d-6a38-4103-ba40-3314a0ccdc48 req-b422675a-e159-4fb1-ba06-3385a5f1a295 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Received event network-vif-plugged-9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 970.419691] env[65788]: DEBUG oslo_concurrency.lockutils [req-1e0d607d-6a38-4103-ba40-3314a0ccdc48 req-b422675a-e159-4fb1-ba06-3385a5f1a295 service nova] Acquiring lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.419950] env[65788]: DEBUG oslo_concurrency.lockutils [req-1e0d607d-6a38-4103-ba40-3314a0ccdc48 req-b422675a-e159-4fb1-ba06-3385a5f1a295 service nova] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.420489] env[65788]: DEBUG oslo_concurrency.lockutils [req-1e0d607d-6a38-4103-ba40-3314a0ccdc48 req-b422675a-e159-4fb1-ba06-3385a5f1a295 service nova] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 970.420652] env[65788]: DEBUG nova.compute.manager [req-1e0d607d-6a38-4103-ba40-3314a0ccdc48 req-b422675a-e159-4fb1-ba06-3385a5f1a295 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] No waiting events found dispatching network-vif-plugged-9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 970.420817] env[65788]: WARNING nova.compute.manager [req-1e0d607d-6a38-4103-ba40-3314a0ccdc48 req-b422675a-e159-4fb1-ba06-3385a5f1a295 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Received unexpected event network-vif-plugged-9fb26674-54b7-46b3-8502-a80cfdfd36aa for instance with vm_state building and task_state spawning. [ 970.433272] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 970.433272] env[65788]: value = "vm-910357" [ 970.433272] env[65788]: _type = "VirtualMachine" [ 970.433272] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 970.433595] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f0461297-e2f5-41c7-a3c4-3acf70acaf14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.443322] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lease: (returnval){ [ 970.443322] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5226a025-6c3c-1c08-e5a7-37c6fb8c70b1" [ 970.443322] env[65788]: _type = "HttpNfcLease" [ 970.443322] env[65788]: } obtained for exporting VM: (result){ [ 970.443322] env[65788]: value = "vm-910357" [ 970.443322] env[65788]: _type = "VirtualMachine" [ 970.443322] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 970.443667] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the lease: (returnval){ [ 970.443667] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5226a025-6c3c-1c08-e5a7-37c6fb8c70b1" [ 970.443667] env[65788]: _type = "HttpNfcLease" [ 970.443667] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 970.451287] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 970.451287] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5226a025-6c3c-1c08-e5a7-37c6fb8c70b1" [ 970.451287] env[65788]: _type = "HttpNfcLease" [ 970.451287] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 970.505704] env[65788]: WARNING neutronclient.v2_0.client [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.506650] env[65788]: WARNING openstack [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.507101] env[65788]: WARNING openstack [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.588033] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.589030] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.600610] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-2cb53fe5-aa58-479e-9090-0d8509cf164e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.600905] env[65788]: DEBUG nova.objects.instance [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'migration_context' on Instance uuid 2cb53fe5-aa58-479e-9090-0d8509cf164e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.676337] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 970.707261] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 970.707261] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 970.707261] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 970.707261] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 970.707261] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 970.707549] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 970.707757] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.707984] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 970.708236] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 970.708460] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 970.708743] env[65788]: DEBUG nova.virt.hardware [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 970.709868] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8970b98-f689-47ea-a379-3aceecae548a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.720082] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6e77d5-50a8-4332-a2dc-601e0d4d4081 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.731205] env[65788]: DEBUG nova.network.neutron [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 970.745418] env[65788]: DEBUG oslo_vmware.api [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662786, 'name': PowerOffVM_Task, 'duration_secs': 0.215343} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.746948] env[65788]: WARNING openstack [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.747470] env[65788]: WARNING openstack [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.754390] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.754627] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.755477] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa433eef-bb56-441b-a072-8513f3c901a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.787142] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.787719] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.838535] env[65788]: DEBUG nova.scheduler.client.report [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 970.842088] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662782, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.866039] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218a9de-db80-f5d4-70d8-aef52be820b7, 'name': SearchDatastore_Task, 'duration_secs': 0.01512} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.866618] env[65788]: INFO nova.compute.manager [-] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Took 1.85 seconds to deallocate network for instance. [ 970.869031] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eca98e53-ccb0-4a5f-a40a-2dbcd6dcbe10 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.877076] env[65788]: WARNING neutronclient.v2_0.client [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.877833] env[65788]: WARNING openstack [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.878285] env[65788]: WARNING openstack [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.889804] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 970.889804] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ad18a1-329c-122b-da15-ac9aebcad8e8" [ 970.889804] env[65788]: _type = "Task" [ 970.889804] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.890183] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.890378] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.890556] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Deleting the datastore file [datastore2] 6215ea56-f99c-42f3-a2d5-f4d31952c8e6 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.890953] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cda9c592-a0d3-4c0a-8af8-802bad0e9928 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.902944] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ad18a1-329c-122b-da15-ac9aebcad8e8, 'name': SearchDatastore_Task, 'duration_secs': 0.015925} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.904520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.905132] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc/22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.905636] env[65788]: DEBUG oslo_vmware.api [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for the task: (returnval){ [ 970.905636] env[65788]: value = "task-4662789" [ 970.905636] env[65788]: _type = "Task" [ 970.905636] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.906778] env[65788]: WARNING neutronclient.v2_0.client [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 970.907513] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 970.907800] env[65788]: WARNING openstack [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 970.915590] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ed69530-28da-47c4-beca-5c9d0cdccee8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.932038] env[65788]: DEBUG oslo_vmware.api [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.933775] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 970.933775] env[65788]: value = "task-4662790" [ 970.933775] env[65788]: _type = "Task" [ 970.933775] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.944396] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662790, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.958271] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 970.958271] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5226a025-6c3c-1c08-e5a7-37c6fb8c70b1" [ 970.958271] env[65788]: _type = "HttpNfcLease" [ 970.958271] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 970.958666] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 970.958666] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5226a025-6c3c-1c08-e5a7-37c6fb8c70b1" [ 970.958666] env[65788]: _type = "HttpNfcLease" [ 970.958666] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 970.959459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7872da0a-6aa5-473a-83a7-fbfbba874a02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.969153] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e6808c-6aca-403c-7376-7ae632822306/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 970.969430] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e6808c-6aca-403c-7376-7ae632822306/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 971.036580] env[65788]: DEBUG nova.network.neutron [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Updated VIF entry in instance network info cache for port ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 971.037161] env[65788]: DEBUG nova.network.neutron [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Updating instance_info_cache with network_info: [{"id": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "address": "fa:16:3e:6c:31:d3", "network": {"id": "6e4a5321-3d4d-4146-ad63-556e8907c3a7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1463785200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a395530cf6f643d6a727bb2f44a77491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec66b0ca-3c", "ovs_interfaceid": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 971.083688] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-26d20d43-2f7a-4d9f-a79d-cec5803d565c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.101025] env[65788]: DEBUG nova.network.neutron [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updating instance_info_cache with network_info: [{"id": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "address": "fa:16:3e:bd:34:66", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fb26674-54", "ovs_interfaceid": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 971.103535] env[65788]: DEBUG nova.objects.base [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Object Instance<2cb53fe5-aa58-479e-9090-0d8509cf164e> lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 971.105167] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944841eb-8114-4c00-a6d6-9cbfb4000f03 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.134521] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13cfc91d-c9e9-40ac-a7f1-35e59ea7e3ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.144311] env[65788]: DEBUG oslo_vmware.api [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 971.144311] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5268fb8d-1460-edf5-924e-30c767a27101" [ 971.144311] env[65788]: _type = "Task" [ 971.144311] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.161470] env[65788]: DEBUG oslo_vmware.api [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5268fb8d-1460-edf5-924e-30c767a27101, 'name': SearchDatastore_Task, 'duration_secs': 0.009294} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.161904] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.342686] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662782, 'name': CreateSnapshot_Task, 'duration_secs': 1.567147} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.343805] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.233s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.347131] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 971.350863] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.061s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.351550] env[65788]: DEBUG nova.objects.instance [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'resources' on Instance uuid f160e4a5-1e91-495e-800e-116ec435d8e1 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.353520] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7224e6-c9c3-489b-97b1-5f98ad5dbfbb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.377557] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.385250] env[65788]: INFO nova.scheduler.client.report [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Deleted allocations for instance 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8 [ 971.430605] env[65788]: DEBUG oslo_vmware.api [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Task: {'id': task-4662789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254258} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.431053] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.431133] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.431316] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.431490] env[65788]: INFO nova.compute.manager [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Took 1.26 seconds to destroy the instance on the hypervisor. [ 971.431761] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 971.432505] env[65788]: DEBUG nova.compute.manager [-] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 971.432604] env[65788]: DEBUG nova.network.neutron [-] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 971.432919] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 971.433496] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 971.433765] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 971.451429] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662790, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.540499] env[65788]: DEBUG oslo_concurrency.lockutils [req-4918e223-be9d-4593-bf53-29bf6cc0524d req-98aa0de8-fb00-44f3-b57b-8221c7e2ef34 service nova] Releasing lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.604409] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 971.606901] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.607336] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Instance network_info: |[{"id": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "address": "fa:16:3e:bd:34:66", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fb26674-54", "ovs_interfaceid": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 971.608155] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:34:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9fb26674-54b7-46b3-8502-a80cfdfd36aa', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.618068] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 971.618149] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.618355] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f9b863d-bf2f-4383-8fb0-929d4e2ea467 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.639808] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.639808] env[65788]: value = "task-4662791" [ 971.639808] env[65788]: _type = "Task" [ 971.639808] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.649160] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662791, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.707515] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 971.707775] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910355', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'name': 'volume-86b8defd-6720-439f-9a6c-b759609a2da3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7f637326-9388-4d73-a1c7-3e4611ac46c5', 'attached_at': '', 'detached_at': '', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'serial': '86b8defd-6720-439f-9a6c-b759609a2da3'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 971.708857] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f81ea0c-ad49-45b4-a5c2-bc89268280b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.729083] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15886514-da0d-4fb3-a652-e1fbb9baa44c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.760111] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] volume-86b8defd-6720-439f-9a6c-b759609a2da3/volume-86b8defd-6720-439f-9a6c-b759609a2da3.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.760702] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff4bd66d-ee34-4ae0-a9f2-c059b254aef4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.782973] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 971.782973] env[65788]: value = "task-4662792" [ 971.782973] env[65788]: _type = "Task" [ 971.782973] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.795172] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662792, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.878591] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 971.879147] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f7925db9-fca0-4c6e-bf25-47615672eb8b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.890610] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 971.890610] env[65788]: value = "task-4662793" [ 971.890610] env[65788]: _type = "Task" [ 971.890610] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.900261] env[65788]: DEBUG oslo_concurrency.lockutils [None req-810acbe6-71d5-4f73-ba4f-fc38c9f5dc89 tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "6a69d645-f89d-43b6-8bdc-b12ce0d88ca8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.405s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.909238] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662793, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.954066] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662790, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660418} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.958569] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc/22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.958850] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.959425] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed7fa7c1-45cd-44dc-af2f-fd3821a92eb4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.970030] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 971.970030] env[65788]: value = "task-4662794" [ 971.970030] env[65788]: _type = "Task" [ 971.970030] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.988296] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662794, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.020378] env[65788]: DEBUG nova.network.neutron [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Successfully updated port: 5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 972.159661] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662791, 'name': CreateVM_Task, 'duration_secs': 0.427942} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.159661] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 972.160165] env[65788]: WARNING neutronclient.v2_0.client [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 972.161086] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.161406] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.161601] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 972.166220] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ddfcd79-276a-440b-adb0-6776054b5fc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.174029] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 972.174029] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526e4813-7036-7e9e-4cd8-94cc1c02b4dc" [ 972.174029] env[65788]: _type = "Task" [ 972.174029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.184632] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526e4813-7036-7e9e-4cd8-94cc1c02b4dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.296590] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662792, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.327424] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e43f27c-65cb-4d09-af7d-ea02abb9fec2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.340825] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59c5266-c859-431b-a1bc-67434f962681 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.386126] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146a5011-4357-4303-87b2-7fcd94d318fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.400226] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855bf476-5908-4ed2-b073-6f2bfa6ad9a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.410871] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662793, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.424029] env[65788]: DEBUG nova.compute.provider_tree [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.481857] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662794, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125058} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.482339] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.483290] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c57c7b-e65a-482b-a268-45a2ed962cb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.511462] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc/22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.512172] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daadab58-ad7a-4e11-b3f1-634dabab8f93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.530030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.530030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.530030] env[65788]: DEBUG nova.network.neutron [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 972.538031] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 972.538031] env[65788]: value = "task-4662795" [ 972.538031] env[65788]: _type = "Task" [ 972.538031] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.551784] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662795, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.656552] env[65788]: DEBUG nova.compute.manager [req-a5979403-649b-4662-9665-a738caa78df4 req-81962b1c-389c-4a77-8f2e-0cb143835dd9 service nova] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Received event network-vif-deleted-a0e489a6-9195-4339-a917-3486d9f0da45 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 972.688539] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526e4813-7036-7e9e-4cd8-94cc1c02b4dc, 'name': SearchDatastore_Task, 'duration_secs': 0.013019} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.688539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 972.688754] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.689872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.689872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.689872] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.690095] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf8538a5-d4e0-435b-9f33-6b913d1b5796 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.703113] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.703458] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.704391] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42453546-a633-4fe9-a811-7a55e1502629 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.712465] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 972.712465] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523002c7-e8bc-5d4c-e984-5bfce1894c36" [ 972.712465] env[65788]: _type = "Task" [ 972.712465] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.723457] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523002c7-e8bc-5d4c-e984-5bfce1894c36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.795934] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662792, 'name': ReconfigVM_Task, 'duration_secs': 0.855726} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.796413] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to attach disk [datastore2] volume-86b8defd-6720-439f-9a6c-b759609a2da3/volume-86b8defd-6720-439f-9a6c-b759609a2da3.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.802579] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0446dfd8-22f7-4b3f-93db-d444ceacc6c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.822381] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 972.822381] env[65788]: value = "task-4662796" [ 972.822381] env[65788]: _type = "Task" [ 972.822381] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.836374] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662796, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.854342] env[65788]: DEBUG nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Received event network-changed-9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 972.854812] env[65788]: DEBUG nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Refreshing instance network info cache due to event network-changed-9fb26674-54b7-46b3-8502-a80cfdfd36aa. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 972.854917] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Acquiring lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.855287] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Acquired lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.855538] env[65788]: DEBUG nova.network.neutron [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Refreshing network info cache for port 9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 972.891055] env[65788]: DEBUG nova.network.neutron [-] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 972.904431] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662793, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.926759] env[65788]: DEBUG nova.scheduler.client.report [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 973.039186] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.040693] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.079162] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662795, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.207024] env[65788]: DEBUG nova.network.neutron [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 973.226680] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523002c7-e8bc-5d4c-e984-5bfce1894c36, 'name': SearchDatastore_Task, 'duration_secs': 0.027148} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.228753] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-287db097-8bb5-4447-b94f-220047a103aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.237055] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 973.237055] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52379a00-d433-32b8-292c-8456742ef9a4" [ 973.237055] env[65788]: _type = "Task" [ 973.237055] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.250957] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52379a00-d433-32b8-292c-8456742ef9a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.279157] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.279698] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.330248] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.330376] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.339413] env[65788]: DEBUG oslo_vmware.api [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662796, 'name': ReconfigVM_Task, 'duration_secs': 0.201904} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.345054] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910355', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'name': 'volume-86b8defd-6720-439f-9a6c-b759609a2da3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '7f637326-9388-4d73-a1c7-3e4611ac46c5', 'attached_at': '', 'detached_at': '', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'serial': '86b8defd-6720-439f-9a6c-b759609a2da3'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 973.358619] env[65788]: WARNING neutronclient.v2_0.client [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.359545] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.359977] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.389815] env[65788]: WARNING neutronclient.v2_0.client [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.391147] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.391579] env[65788]: WARNING openstack [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.400502] env[65788]: INFO nova.compute.manager [-] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Took 1.97 seconds to deallocate network for instance. [ 973.420108] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662793, 'name': CloneVM_Task, 'duration_secs': 1.476225} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.420501] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Created linked-clone VM from snapshot [ 973.421853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3008be8-0ba1-4ce1-874c-50be73cdc107 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.433506] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.082s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.441515] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.402s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 973.442036] env[65788]: DEBUG nova.objects.instance [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lazy-loading 'resources' on Instance uuid e0f7c849-315c-4247-a840-d388d48746b9 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.443438] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Uploading image 3710be11-4cef-4377-96f6-97f944468c82 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 973.478171] env[65788]: INFO nova.scheduler.client.report [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted allocations for instance f160e4a5-1e91-495e-800e-116ec435d8e1 [ 973.573328] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662795, 'name': ReconfigVM_Task, 'duration_secs': 0.56599} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.574165] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc/22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.575406] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71f7ac7c-754f-4b1a-8264-4a9db879aedc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.582638] env[65788]: DEBUG nova.network.neutron [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updating instance_info_cache with network_info: [{"id": "5c238645-7548-466b-983b-2ad80497b15e", "address": "fa:16:3e:38:90:25", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c238645-75", "ovs_interfaceid": "5c238645-7548-466b-983b-2ad80497b15e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 973.587921] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 973.587921] env[65788]: value = "task-4662797" [ 973.587921] env[65788]: _type = "Task" [ 973.587921] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.600478] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662797, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.669190] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.669740] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.760957] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52379a00-d433-32b8-292c-8456742ef9a4, 'name': SearchDatastore_Task, 'duration_secs': 0.015659} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.761746] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 973.761746] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 29c0ab17-5df2-4f6c-bdea-3a4a6c857066/29c0ab17-5df2-4f6c-bdea-3a4a6c857066.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.762207] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54e9f329-ebae-46ac-a3fe-6d87bed9fbc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.775340] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 973.775340] env[65788]: value = "task-4662798" [ 973.775340] env[65788]: _type = "Task" [ 973.775340] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.789361] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.801908] env[65788]: WARNING neutronclient.v2_0.client [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 973.802900] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 973.803482] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 973.841145] env[65788]: INFO nova.compute.manager [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Detaching volume a5111781-1b06-4e45-b9ac-40d76784a6cc [ 973.898274] env[65788]: INFO nova.virt.block_device [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Attempting to driver detach volume a5111781-1b06-4e45-b9ac-40d76784a6cc from mountpoint /dev/sdb [ 973.898274] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 973.898274] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910320', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'name': 'volume-a5111781-1b06-4e45-b9ac-40d76784a6cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3a06a833-5aaa-4b5d-88b3-8a1d469580af', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'serial': 'a5111781-1b06-4e45-b9ac-40d76784a6cc'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 973.899308] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20feed8-3ff1-4d0c-9320-f6b484d50865 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.931351] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.938843] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e4a0cc-41e3-4c40-8692-b041f114df0b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.958015] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99772e1-8462-415f-827d-36e95b534d63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.989163] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8a7c09fa-046f-44de-9ec3-d0b026ff51de tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "f160e4a5-1e91-495e-800e-116ec435d8e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.951s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 973.994239] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e0e175-51c0-4301-9ae3-9760c43dadbd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.013987] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] The volume has not been displaced from its original location: [datastore2] volume-a5111781-1b06-4e45-b9ac-40d76784a6cc/volume-a5111781-1b06-4e45-b9ac-40d76784a6cc.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 974.019996] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfiguring VM instance instance-0000003b to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 974.022820] env[65788]: DEBUG nova.network.neutron [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updated VIF entry in instance network info cache for port 9fb26674-54b7-46b3-8502-a80cfdfd36aa. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 974.023259] env[65788]: DEBUG nova.network.neutron [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updating instance_info_cache with network_info: [{"id": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "address": "fa:16:3e:bd:34:66", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fb26674-54", "ovs_interfaceid": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 974.025838] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04348b5e-08cc-4074-a412-3e8d2312a88c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.047305] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Releasing lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.047684] env[65788]: DEBUG nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Received event network-vif-plugged-5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 974.048352] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Acquiring lock "af9120a6-2cea-46da-ba06-6036b9beda78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 974.048352] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Lock "af9120a6-2cea-46da-ba06-6036b9beda78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.048470] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Lock "af9120a6-2cea-46da-ba06-6036b9beda78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.048858] env[65788]: DEBUG nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] No waiting events found dispatching network-vif-plugged-5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 974.048946] env[65788]: WARNING nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Received unexpected event network-vif-plugged-5c238645-7548-466b-983b-2ad80497b15e for instance with vm_state building and task_state spawning. [ 974.049236] env[65788]: DEBUG nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Received event network-changed-5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 974.049446] env[65788]: DEBUG nova.compute.manager [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Refreshing instance network info cache due to event network-changed-5c238645-7548-466b-983b-2ad80497b15e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 974.049672] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Acquiring lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.059509] env[65788]: DEBUG oslo_vmware.api [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 974.059509] env[65788]: value = "task-4662799" [ 974.059509] env[65788]: _type = "Task" [ 974.059509] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.072954] env[65788]: DEBUG oslo_vmware.api [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662799, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.093034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.093887] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Instance network_info: |[{"id": "5c238645-7548-466b-983b-2ad80497b15e", "address": "fa:16:3e:38:90:25", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c238645-75", "ovs_interfaceid": "5c238645-7548-466b-983b-2ad80497b15e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 974.095124] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Acquired lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 974.095579] env[65788]: DEBUG nova.network.neutron [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Refreshing network info cache for port 5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 974.097571] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:90:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604056d6-6dd6-47fa-9eaa-6863a3a7c488', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c238645-7548-466b-983b-2ad80497b15e', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.107863] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 974.114261] env[65788]: WARNING neutronclient.v2_0.client [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 974.114993] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.115378] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.122576] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.126302] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7f6ee1a-2272-4751-b32d-1ae3c133afc4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.151686] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662797, 'name': Rename_Task, 'duration_secs': 0.234588} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.154589] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.154886] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.154886] env[65788]: value = "task-4662800" [ 974.154886] env[65788]: _type = "Task" [ 974.154886] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.155327] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-951c0c41-dc15-4ffd-8420-7bb367a24aa8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.171391] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662800, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.176332] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 974.176332] env[65788]: value = "task-4662801" [ 974.176332] env[65788]: _type = "Task" [ 974.176332] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.293273] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.294141] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.312417] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662798, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.407725] env[65788]: DEBUG nova.objects.instance [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'flavor' on Instance uuid 7f637326-9388-4d73-a1c7-3e4611ac46c5 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.411113] env[65788]: WARNING neutronclient.v2_0.client [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 974.412147] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 974.412606] env[65788]: WARNING openstack [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 974.464693] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ebf7ea-c827-4745-ae84-5eb13feca76e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.476325] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bfbc40-18dc-4700-8843-d0cb71f3830f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.541599] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b10bde1-dbf1-4503-ae8b-bbde2e760372 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.555756] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57aa2b7-a136-4b9f-a022-59432be99509 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.581190] env[65788]: DEBUG nova.compute.provider_tree [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.583054] env[65788]: DEBUG oslo_vmware.api [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662799, 'name': ReconfigVM_Task, 'duration_secs': 0.32983} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.584031] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Reconfigured VM instance instance-0000003b to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 974.591096] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3e48906-6283-4b2d-8e2b-b4004f4e6d69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.608331] env[65788]: DEBUG oslo_vmware.api [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 974.608331] env[65788]: value = "task-4662802" [ 974.608331] env[65788]: _type = "Task" [ 974.608331] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.622293] env[65788]: DEBUG oslo_vmware.api [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662802, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.671164] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662800, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.692633] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662801, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.701977] env[65788]: DEBUG nova.network.neutron [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updated VIF entry in instance network info cache for port 5c238645-7548-466b-983b-2ad80497b15e. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 974.702133] env[65788]: DEBUG nova.network.neutron [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updating instance_info_cache with network_info: [{"id": "5c238645-7548-466b-983b-2ad80497b15e", "address": "fa:16:3e:38:90:25", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c238645-75", "ovs_interfaceid": "5c238645-7548-466b-983b-2ad80497b15e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 974.796297] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.842953} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.796825] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 29c0ab17-5df2-4f6c-bdea-3a4a6c857066/29c0ab17-5df2-4f6c-bdea-3a4a6c857066.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.796825] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.797161] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cc81fb56-14fe-4293-8387-a8aa83fb28f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.807573] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 974.807573] env[65788]: value = "task-4662803" [ 974.807573] env[65788]: _type = "Task" [ 974.807573] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.822166] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.896225] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 974.896643] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-55cef79f-e314-4e87-8ed0-b8cf02002cba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.906745] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 974.906745] env[65788]: value = "task-4662804" [ 974.906745] env[65788]: _type = "Task" [ 974.906745] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.927362] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662804, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.928811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7ca147a2-7e81-4605-beb0-05d828b4a1c9 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.358s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 974.931581] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5274a35e-6215-d418-e069-b7938364d500/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 974.932330] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddc88f7-112f-4b53-9aa0-65e36576e5f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.944332] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5274a35e-6215-d418-e069-b7938364d500/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 974.944332] env[65788]: ERROR oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5274a35e-6215-d418-e069-b7938364d500/disk-0.vmdk due to incomplete transfer. [ 974.944661] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e05df767-888f-4d57-98c3-61b84e29ef39 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.954768] env[65788]: DEBUG oslo_vmware.rw_handles [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5274a35e-6215-d418-e069-b7938364d500/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 974.955140] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Uploaded image 0b0e4bc3-60eb-4ac5-b652-2cc27f204568 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 974.959187] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 974.962980] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8518c48d-2efd-4540-a0e5-9988f16ce29a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.972540] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 974.972540] env[65788]: value = "task-4662805" [ 974.972540] env[65788]: _type = "Task" [ 974.972540] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.983982] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662805, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.085334] env[65788]: DEBUG nova.scheduler.client.report [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 975.101793] env[65788]: DEBUG nova.compute.manager [req-7abeaee4-fcc0-40cf-bb34-34b07eb9367d req-8eddd91b-e160-4460-94b9-1fc2ee613967 service nova] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Received event network-vif-deleted-fcafb23d-cf58-408a-86ee-d14ae37c1bda {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 975.120315] env[65788]: DEBUG oslo_vmware.api [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662802, 'name': ReconfigVM_Task, 'duration_secs': 0.18583} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.120611] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910320', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'name': 'volume-a5111781-1b06-4e45-b9ac-40d76784a6cc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3a06a833-5aaa-4b5d-88b3-8a1d469580af', 'attached_at': '', 'detached_at': '', 'volume_id': 'a5111781-1b06-4e45-b9ac-40d76784a6cc', 'serial': 'a5111781-1b06-4e45-b9ac-40d76784a6cc'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 975.171038] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662800, 'name': CreateVM_Task, 'duration_secs': 0.836475} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.171038] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.172052] env[65788]: WARNING neutronclient.v2_0.client [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 975.172052] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.172155] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.172772] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 975.173078] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ee00212-1de1-4572-b322-9cda3c40d3d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.183188] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 975.183188] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5217ba56-a5d6-12c1-ecde-70d78f41a314" [ 975.183188] env[65788]: _type = "Task" [ 975.183188] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.191658] env[65788]: DEBUG oslo_vmware.api [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662801, 'name': PowerOnVM_Task, 'duration_secs': 0.910504} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.192445] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.192664] env[65788]: INFO nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Took 9.62 seconds to spawn the instance on the hypervisor. [ 975.192881] env[65788]: DEBUG nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 975.193870] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7860c19f-90aa-4b94-92e2-690533b6a412 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.201923] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5217ba56-a5d6-12c1-ecde-70d78f41a314, 'name': SearchDatastore_Task, 'duration_secs': 0.014024} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.202696] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 975.202927] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.203130] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.203250] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 975.203448] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.203767] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b349f29-0c87-4a42-89dd-336bab676d8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.207816] env[65788]: DEBUG oslo_concurrency.lockutils [req-1ef5f15f-959f-4b8d-ab84-2ec5215e3ef0 req-7cadc64f-a9b4-4539-9cdc-0a6fe68f0134 service nova] Releasing lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 975.220082] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.220311] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 975.221359] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d14f219-95a1-4df1-aff5-741edbcb2047 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.230727] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 975.230727] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52989b4d-2646-201e-8654-04278b6e3fb2" [ 975.230727] env[65788]: _type = "Task" [ 975.230727] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.240948] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52989b4d-2646-201e-8654-04278b6e3fb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.319340] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143661} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.319628] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.320507] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f9608f-0591-47ab-b749-2008cc2b8a9a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.357212] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 29c0ab17-5df2-4f6c-bdea-3a4a6c857066/29c0ab17-5df2-4f6c-bdea-3a4a6c857066.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.357447] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceb0d675-e309-4656-a69d-881881c28787 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.389721] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 975.389721] env[65788]: value = "task-4662806" [ 975.389721] env[65788]: _type = "Task" [ 975.389721] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.400542] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662806, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.422470] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662804, 'name': Destroy_Task, 'duration_secs': 0.511232} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.422948] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Destroyed the VM [ 975.423283] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 975.423569] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9cd071e5-5ee2-48ba-8cca-534dd6938440 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.432311] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 975.432311] env[65788]: value = "task-4662807" [ 975.432311] env[65788]: _type = "Task" [ 975.432311] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.442436] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662807, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.486360] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662805, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.592399] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.151s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.595699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.434s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.616479] env[65788]: INFO nova.scheduler.client.report [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleted allocations for instance e0f7c849-315c-4247-a840-d388d48746b9 [ 975.686218] env[65788]: DEBUG nova.objects.instance [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'flavor' on Instance uuid 3a06a833-5aaa-4b5d-88b3-8a1d469580af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.721463] env[65788]: INFO nova.compute.manager [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Took 30.23 seconds to build instance. [ 975.742757] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52989b4d-2646-201e-8654-04278b6e3fb2, 'name': SearchDatastore_Task, 'duration_secs': 0.01168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.744154] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab2697bf-63cb-4831-9499-160737dd744d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.751745] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 975.751745] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b61482-c903-917c-9997-0f96337afaf0" [ 975.751745] env[65788]: _type = "Task" [ 975.751745] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.766083] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b61482-c903-917c-9997-0f96337afaf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.904235] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662806, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.952840] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662807, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.985248] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662805, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.129486] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6607bd8f-eabd-41be-b50c-010db4b40067 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "e0f7c849-315c-4247-a840-d388d48746b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.644s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.224637] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eab74e8d-9b07-4f4d-96f6-94b588c08d13 tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.757s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.266436] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b61482-c903-917c-9997-0f96337afaf0, 'name': SearchDatastore_Task, 'duration_secs': 0.031552} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.269759] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 976.270059] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] af9120a6-2cea-46da-ba06-6036b9beda78/af9120a6-2cea-46da-ba06-6036b9beda78.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.271657] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f635b17e-7637-4027-a7db-04c1178e0e1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.280076] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 976.280076] env[65788]: value = "task-4662808" [ 976.280076] env[65788]: _type = "Task" [ 976.280076] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.294412] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.402280] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662806, 'name': ReconfigVM_Task, 'duration_secs': 0.595181} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.405281] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 29c0ab17-5df2-4f6c-bdea-3a4a6c857066/29c0ab17-5df2-4f6c-bdea-3a4a6c857066.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 976.406214] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07394078-8fab-47d0-997e-8883c0582846 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.414626] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 976.414626] env[65788]: value = "task-4662809" [ 976.414626] env[65788]: _type = "Task" [ 976.414626] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.429548] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662809, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.439919] env[65788]: DEBUG nova.compute.manager [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 976.443895] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.444264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.461617] env[65788]: DEBUG oslo_vmware.api [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662807, 'name': RemoveSnapshot_Task, 'duration_secs': 0.798045} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.463653] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 976.466529] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7ec10c-3fed-49f4-932b-154804049669 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.479980] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5fbe62-4a98-4123-b752-386dbdb8a2c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.490799] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662805, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.517416] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909c8e13-a698-4263-817e-e01c50d136c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.527490] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f173018f-1b17-4377-a42a-0e4dcac9a508 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.550503] env[65788]: DEBUG nova.compute.provider_tree [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.696748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa71ccc-896f-460a-8453-6135994db466 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.366s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 976.796036] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662808, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.932429] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662809, 'name': Rename_Task, 'duration_secs': 0.237419} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.932929] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.933479] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22aba6e4-6bab-43f7-8bcd-2ac39701ba66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.943086] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 976.943086] env[65788]: value = "task-4662810" [ 976.943086] env[65788]: _type = "Task" [ 976.943086] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.955396] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 976.963729] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.970397] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.971667] env[65788]: WARNING nova.compute.manager [None req-04046254-f5ec-4b4c-8c98-52ce21b3c0f7 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Image not found during snapshot: nova.exception.ImageNotFound: Image 3710be11-4cef-4377-96f6-97f944468c82 could not be found. [ 976.986956] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662805, 'name': Destroy_Task, 'duration_secs': 1.725534} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.986956] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Destroyed the VM [ 976.987169] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 976.988154] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6a1dab6c-6292-498c-9142-c254fb7d4d1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.995728] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 976.995728] env[65788]: value = "task-4662811" [ 976.995728] env[65788]: _type = "Task" [ 976.995728] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.007180] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662811, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.056423] env[65788]: DEBUG nova.scheduler.client.report [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.066593] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "1a701758-a056-4948-9069-2a7168f2cc8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.067182] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "1a701758-a056-4948-9069-2a7168f2cc8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.294252] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662808, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.711918} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.294732] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] af9120a6-2cea-46da-ba06-6036b9beda78/af9120a6-2cea-46da-ba06-6036b9beda78.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.294992] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.295314] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c261cff-0d6d-4a95-a4ba-aaefb120d415 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.309810] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 977.309810] env[65788]: value = "task-4662812" [ 977.309810] env[65788]: _type = "Task" [ 977.309810] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.325407] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.458419] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662810, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.489766] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.499030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.499030] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.510922] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662811, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.569756] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 977.598930] env[65788]: DEBUG nova.compute.manager [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Received event network-changed-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 977.599240] env[65788]: DEBUG nova.compute.manager [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Refreshing instance network info cache due to event network-changed-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 977.599826] env[65788]: DEBUG oslo_concurrency.lockutils [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Acquiring lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.600083] env[65788]: DEBUG oslo_concurrency.lockutils [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Acquired lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 977.600389] env[65788]: DEBUG nova.network.neutron [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Refreshing network info cache for port ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 977.646912] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.647065] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.647252] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 977.647482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 977.647692] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.655045] env[65788]: INFO nova.compute.manager [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Terminating instance [ 977.824496] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.414007} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.824496] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.824966] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f5a761-dd07-439c-80ac-66c0cd6a50eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.852933] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] af9120a6-2cea-46da-ba06-6036b9beda78/af9120a6-2cea-46da-ba06-6036b9beda78.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.853367] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45ea5d97-538b-4e3d-b2bf-27ecd7aa1e6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.877876] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 977.877876] env[65788]: value = "task-4662813" [ 977.877876] env[65788]: _type = "Task" [ 977.877876] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.887992] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662813, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.960836] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662810, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.999760] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 978.013415] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662811, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.067354] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.471s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.071552] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.694s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.071920] env[65788]: DEBUG nova.objects.instance [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lazy-loading 'resources' on Instance uuid a79bbb7e-7b78-4900-9b83-bfce94bb41bd {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.104111] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.104929] env[65788]: WARNING neutronclient.v2_0.client [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 978.105568] env[65788]: WARNING openstack [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.105923] env[65788]: WARNING openstack [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.162777] env[65788]: DEBUG nova.compute.manager [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 978.163137] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.164294] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd95a17-1d22-4098-a225-e2a2492c67bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.174910] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.175935] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c38e9260-b970-4ee0-8d6d-2448a3a20581 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.185088] env[65788]: DEBUG oslo_vmware.api [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 978.185088] env[65788]: value = "task-4662814" [ 978.185088] env[65788]: _type = "Task" [ 978.185088] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.189708] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.190015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.190301] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.190449] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.192053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.193429] env[65788]: INFO nova.compute.manager [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Terminating instance [ 978.204969] env[65788]: DEBUG oslo_vmware.api [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.274264] env[65788]: WARNING openstack [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.274264] env[65788]: WARNING openstack [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.359674] env[65788]: WARNING neutronclient.v2_0.client [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 978.360539] env[65788]: WARNING openstack [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 978.361133] env[65788]: WARNING openstack [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 978.391944] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662813, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.457726] env[65788]: DEBUG oslo_vmware.api [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662810, 'name': PowerOnVM_Task, 'duration_secs': 1.353465} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.458156] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.458627] env[65788]: INFO nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Took 9.85 seconds to spawn the instance on the hypervisor. [ 978.458627] env[65788]: DEBUG nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 978.459385] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47068a11-0fa9-4d52-a35b-39726f236721 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.472446] env[65788]: DEBUG nova.network.neutron [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Updated VIF entry in instance network info cache for port ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 978.472858] env[65788]: DEBUG nova.network.neutron [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Updating instance_info_cache with network_info: [{"id": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "address": "fa:16:3e:6c:31:d3", "network": {"id": "6e4a5321-3d4d-4146-ad63-556e8907c3a7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1463785200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a395530cf6f643d6a727bb2f44a77491", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec66b0ca-3c", "ovs_interfaceid": "ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 978.513271] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662811, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.532165] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.633142] env[65788]: INFO nova.scheduler.client.report [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocation for migration 8b1f0895-2e84-4d39-9589-aa78a5a0a9b1 [ 978.699357] env[65788]: DEBUG nova.compute.manager [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 978.699584] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.699895] env[65788]: DEBUG oslo_vmware.api [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662814, 'name': PowerOffVM_Task, 'duration_secs': 0.308922} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.704996] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8627b77a-78fa-4c13-8398-a3b330b0e08e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.708590] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.708923] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.709609] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af07aeb8-6c58-47e2-96e9-ce0f42d57ec2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.718617] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.719401] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-709341df-00e2-44aa-b857-072eb08fe303 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.732695] env[65788]: DEBUG oslo_vmware.api [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 978.732695] env[65788]: value = "task-4662816" [ 978.732695] env[65788]: _type = "Task" [ 978.732695] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.751527] env[65788]: DEBUG oslo_vmware.api [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662816, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.813118] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.815213] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.815213] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleting the datastore file [datastore2] 3a06a833-5aaa-4b5d-88b3-8a1d469580af {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.815213] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f536da5f-0434-47dc-98a4-d3793d5cc7e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.825615] env[65788]: DEBUG oslo_vmware.api [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 978.825615] env[65788]: value = "task-4662817" [ 978.825615] env[65788]: _type = "Task" [ 978.825615] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.840879] env[65788]: DEBUG oslo_vmware.api [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.891366] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662813, 'name': ReconfigVM_Task, 'duration_secs': 0.569751} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.891366] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Reconfigured VM instance instance-00000054 to attach disk [datastore2] af9120a6-2cea-46da-ba06-6036b9beda78/af9120a6-2cea-46da-ba06-6036b9beda78.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.891526] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c7ba87c-b1da-4e47-911b-00333089b211 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.904401] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 978.904401] env[65788]: value = "task-4662818" [ 978.904401] env[65788]: _type = "Task" [ 978.904401] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.917203] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662818, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.960252] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20fb840b-6d5c-4a02-9633-5f19350fb3d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.970645] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff809e28-9ed5-45d6-a64d-599e8443e241 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.979385] env[65788]: DEBUG oslo_concurrency.lockutils [req-ccefa82b-830c-4852-bc03-5a4e28d6313c req-f1bca5a1-fb97-47f9-bb1d-d5c4dc138158 service nova] Releasing lock "refresh_cache-22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 979.013409] env[65788]: INFO nova.compute.manager [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Took 27.84 seconds to build instance. [ 979.018502] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b98dc7-70f0-4795-a4f1-9ca03b32158e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.034526] env[65788]: DEBUG oslo_vmware.api [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662811, 'name': RemoveSnapshot_Task, 'duration_secs': 1.810194} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.035052] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 979.035446] env[65788]: INFO nova.compute.manager [None req-e97c8ca8-2613-408d-a334-a76a8f72d165 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Took 18.25 seconds to snapshot the instance on the hypervisor. [ 979.039723] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d3b550-f99c-48a6-9981-fc4f612d57ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.057731] env[65788]: DEBUG nova.compute.provider_tree [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.143154] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad7c38b7-10ce-4ad9-a581-2468fbf103f7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 12.578s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.250912] env[65788]: DEBUG oslo_vmware.api [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662816, 'name': PowerOffVM_Task, 'duration_secs': 0.427489} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.251254] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.251461] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.251765] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6312f8dd-6d4f-49fb-a7cc-c958daf9d426 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.333615] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.333780] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.333888] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleting the datastore file [datastore2] b220df77-cd3a-4aeb-87ad-68f56995ec3b {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.337231] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5386a5e-9362-4e70-ae44-bbe28d4f2f35 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.339409] env[65788]: DEBUG oslo_vmware.api [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280225} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.339658] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.339838] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.341234] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.341234] env[65788]: INFO nova.compute.manager [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Took 1.18 seconds to destroy the instance on the hypervisor. [ 979.341234] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 979.341234] env[65788]: DEBUG nova.compute.manager [-] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 979.341602] env[65788]: DEBUG nova.network.neutron [-] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 979.341602] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.342101] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 979.342374] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 979.350962] env[65788]: DEBUG oslo_vmware.api [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for the task: (returnval){ [ 979.350962] env[65788]: value = "task-4662820" [ 979.350962] env[65788]: _type = "Task" [ 979.350962] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.360352] env[65788]: DEBUG oslo_vmware.api [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662820, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.385442] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.416650] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662818, 'name': Rename_Task, 'duration_secs': 0.243556} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.416958] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.417263] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f7002ad-1ea0-4381-bcca-f116c3eab431 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.427416] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 979.427416] env[65788]: value = "task-4662821" [ 979.427416] env[65788]: _type = "Task" [ 979.427416] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.436876] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662821, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.522621] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a51723d-baba-46fc-b671-5fe6f5053190 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.382s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.561105] env[65788]: DEBUG nova.scheduler.client.report [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.863541] env[65788]: DEBUG oslo_vmware.api [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Task: {'id': task-4662820, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193368} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.863903] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.864069] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.864204] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.864529] env[65788]: INFO nova.compute.manager [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Took 1.16 seconds to destroy the instance on the hypervisor. [ 979.864644] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 979.864833] env[65788]: DEBUG nova.compute.manager [-] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 979.864943] env[65788]: DEBUG nova.network.neutron [-] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 979.865708] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.865874] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 979.866132] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 979.893453] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.893837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.894151] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 979.894420] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 979.894666] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.901332] env[65788]: INFO nova.compute.manager [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Terminating instance [ 979.924754] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 979.939167] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662821, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.070925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.999s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.074157] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.143s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.074662] env[65788]: DEBUG nova.objects.instance [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lazy-loading 'resources' on Instance uuid 6215ea56-f99c-42f3-a2d5-f4d31952c8e6 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.106040] env[65788]: INFO nova.scheduler.client.report [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Deleted allocations for instance a79bbb7e-7b78-4900-9b83-bfce94bb41bd [ 980.347661] env[65788]: DEBUG nova.network.neutron [-] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.407209] env[65788]: DEBUG nova.compute.manager [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 980.408230] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 980.408555] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8076707f-8541-442f-bc23-85ba3329f41a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.417323] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 980.417592] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c623997a-0c38-436d-bd9e-997055a1629b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.424305] env[65788]: DEBUG oslo_vmware.api [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 980.424305] env[65788]: value = "task-4662822" [ 980.424305] env[65788]: _type = "Task" [ 980.424305] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.437321] env[65788]: DEBUG oslo_vmware.api [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.439965] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e6808c-6aca-403c-7376-7ae632822306/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 980.441736] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d9cce7-8195-417e-b480-41bc2c8e0511 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.447212] env[65788]: DEBUG oslo_vmware.api [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662821, 'name': PowerOnVM_Task, 'duration_secs': 0.608076} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.447910] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.448169] env[65788]: INFO nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Took 9.77 seconds to spawn the instance on the hypervisor. [ 980.448370] env[65788]: DEBUG nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 980.449258] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994a3fa4-69aa-4941-bb03-11208c9cf01a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.454028] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e6808c-6aca-403c-7376-7ae632822306/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 980.454203] env[65788]: ERROR oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e6808c-6aca-403c-7376-7ae632822306/disk-0.vmdk due to incomplete transfer. [ 980.454786] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cbeb40ee-01f7-4944-9d69-830a980399f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.463162] env[65788]: DEBUG oslo_vmware.rw_handles [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e6808c-6aca-403c-7376-7ae632822306/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 980.463235] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Uploaded image 0f208f1a-1c47-4d89-ac7d-a4f584ec31ed to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 980.465225] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 980.465821] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 980.466055] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 980.466229] env[65788]: INFO nova.compute.manager [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Shelving [ 980.467412] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b3c890ac-f35d-4b3e-9a88-13b47b80cec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.476105] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 980.476105] env[65788]: value = "task-4662823" [ 980.476105] env[65788]: _type = "Task" [ 980.476105] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.486966] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662823, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.489167] env[65788]: DEBUG nova.compute.manager [req-f0456f74-933d-4ca9-8088-7bd3282d9ffe req-e6f47667-5ea5-4465-8bea-09da933927ea service nova] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Received event network-vif-deleted-27626e1a-51fe-4f5a-9b93-8ab4c1b8b694 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 980.616793] env[65788]: DEBUG oslo_concurrency.lockutils [None req-55e16580-e798-4dfc-8e40-64faa2ff1d1d tempest-InstanceActionsV221TestJSON-2011935552 tempest-InstanceActionsV221TestJSON-2011935552-project-member] Lock "a79bbb7e-7b78-4900-9b83-bfce94bb41bd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.258s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 980.703045] env[65788]: DEBUG nova.network.neutron [-] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 980.852543] env[65788]: INFO nova.compute.manager [-] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Took 1.51 seconds to deallocate network for instance. [ 980.920308] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e30f7c-7c8e-4e22-b821-bbae6bbc47db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.934215] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2c06ec-8d3b-427d-b80a-d43ecdc6f453 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.946477] env[65788]: DEBUG oslo_vmware.api [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662822, 'name': PowerOffVM_Task, 'duration_secs': 0.371342} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.973015] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 980.973015] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 980.985483] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1119015-0424-4bb2-a395-88a1a1c619ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.987810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada11efb-b9b5-4d56-bf7f-ae3a2d21d61e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.991751] env[65788]: INFO nova.compute.manager [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Took 27.18 seconds to build instance. [ 981.005835] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662823, 'name': Destroy_Task, 'duration_secs': 0.384496} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.006945] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Destroyed the VM [ 981.007362] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 981.008514] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949ff469-d8e8-4191-b6cd-f3fce0fc18bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.012968] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-370452f5-737e-48f8-8ddc-68fc69810008 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.028196] env[65788]: DEBUG nova.compute.provider_tree [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.031451] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 981.031451] env[65788]: value = "task-4662825" [ 981.031451] env[65788]: _type = "Task" [ 981.031451] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.042115] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662825, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.078133] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.078447] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.078685] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore2] 2cb53fe5-aa58-479e-9090-0d8509cf164e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.079021] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a15e23b-f82b-4bb2-9c50-a33490001cb3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.087813] env[65788]: DEBUG oslo_vmware.api [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 981.087813] env[65788]: value = "task-4662826" [ 981.087813] env[65788]: _type = "Task" [ 981.087813] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.098061] env[65788]: DEBUG oslo_vmware.api [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.204972] env[65788]: INFO nova.compute.manager [-] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Took 1.34 seconds to deallocate network for instance. [ 981.361292] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.411538] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "902d8a69-0398-4752-a609-0ac5b4b32e27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.411863] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.486395] env[65788]: DEBUG nova.compute.manager [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Received event network-changed-5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 981.486601] env[65788]: DEBUG nova.compute.manager [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Refreshing instance network info cache due to event network-changed-5c238645-7548-466b-983b-2ad80497b15e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 981.486819] env[65788]: DEBUG oslo_concurrency.lockutils [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Acquiring lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.486961] env[65788]: DEBUG oslo_concurrency.lockutils [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Acquired lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 981.487681] env[65788]: DEBUG nova.network.neutron [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Refreshing network info cache for port 5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 981.496183] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.496470] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-434d2e7e-0ccf-4316-9c14-c0c27fb5ac2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.503655] env[65788]: DEBUG oslo_concurrency.lockutils [None req-752c8886-8264-4cce-a0a4-d6bc4ad7a2c8 tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.705s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.505743] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 981.505743] env[65788]: value = "task-4662827" [ 981.505743] env[65788]: _type = "Task" [ 981.505743] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.516237] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.533731] env[65788]: DEBUG nova.scheduler.client.report [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 981.548480] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662825, 'name': RemoveSnapshot_Task, 'duration_secs': 0.392426} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.550480] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 981.550480] env[65788]: DEBUG nova.compute.manager [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 981.550480] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249569b5-a5ff-4343-bec9-9eb5df9ac6f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.598828] env[65788]: DEBUG oslo_vmware.api [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218265} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.599118] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.599304] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.599622] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.599666] env[65788]: INFO nova.compute.manager [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Took 1.19 seconds to destroy the instance on the hypervisor. [ 981.599891] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 981.600827] env[65788]: DEBUG nova.compute.manager [-] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 981.600907] env[65788]: DEBUG nova.network.neutron [-] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 981.601288] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 981.601715] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 981.601982] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 981.647030] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 981.712928] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 981.914815] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 981.993843] env[65788]: WARNING neutronclient.v2_0.client [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 981.993843] env[65788]: WARNING openstack [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 981.994451] env[65788]: WARNING openstack [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.016912] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662827, 'name': PowerOffVM_Task, 'duration_secs': 0.40464} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.017120] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.018070] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b153ace-4d71-449f-90e3-5807e53af79a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.040475] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca7d722-d490-45f5-8bff-24d4bfb6e95b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.043890] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.046515] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.076s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.071357] env[65788]: INFO nova.compute.manager [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Shelve offloading [ 982.081938] env[65788]: INFO nova.scheduler.client.report [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Deleted allocations for instance 6215ea56-f99c-42f3-a2d5-f4d31952c8e6 [ 982.208324] env[65788]: WARNING openstack [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 982.208856] env[65788]: WARNING openstack [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.253114] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "af9120a6-2cea-46da-ba06-6036b9beda78" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.253437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.253707] env[65788]: INFO nova.compute.manager [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Rebooting instance [ 982.303180] env[65788]: WARNING neutronclient.v2_0.client [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 982.303867] env[65788]: WARNING openstack [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 982.304241] env[65788]: WARNING openstack [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 982.391070] env[65788]: DEBUG nova.network.neutron [-] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 982.448871] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.499097] env[65788]: DEBUG nova.network.neutron [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updated VIF entry in instance network info cache for port 5c238645-7548-466b-983b-2ad80497b15e. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 982.499097] env[65788]: DEBUG nova.network.neutron [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updating instance_info_cache with network_info: [{"id": "5c238645-7548-466b-983b-2ad80497b15e", "address": "fa:16:3e:38:90:25", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c238645-75", "ovs_interfaceid": "5c238645-7548-466b-983b-2ad80497b15e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 982.519453] env[65788]: DEBUG nova.compute.manager [req-15e45582-7263-4d9b-8ddb-fbd8c94acff4 req-0b8fce43-701c-456a-b2e3-045bc31b7aaf service nova] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Received event network-vif-deleted-f5aec569-cc68-4c71-a4ca-10710acfbd44 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 982.551783] env[65788]: INFO nova.compute.claims [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.558051] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 982.558281] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f12f36fc-7976-4731-b8ad-dbef6c1f7d78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.567178] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 982.567178] env[65788]: value = "task-4662828" [ 982.567178] env[65788]: _type = "Task" [ 982.567178] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.579180] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.579502] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662828, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.579730] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00b8afcf-64d6-4d85-9804-2f9ed5f37da3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.588962] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 982.588962] env[65788]: value = "task-4662829" [ 982.588962] env[65788]: _type = "Task" [ 982.588962] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.590271] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2fecc1b9-bf00-405e-bba0-dce79dd76478 tempest-ServersV294TestFqdnHostnames-1151149332 tempest-ServersV294TestFqdnHostnames-1151149332-project-member] Lock "6215ea56-f99c-42f3-a2d5-f4d31952c8e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.922s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.600890] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 982.601152] env[65788]: DEBUG nova.compute.manager [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 982.601942] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c344e064-416a-44a8-99ff-ac82b38b7eac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.608792] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.608964] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 982.609151] env[65788]: DEBUG nova.network.neutron [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 982.776436] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.894454] env[65788]: INFO nova.compute.manager [-] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Took 1.29 seconds to deallocate network for instance. [ 982.951371] env[65788]: DEBUG nova.compute.manager [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 982.952555] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b2dbef-a604-40ef-9328-56252d4420e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.001732] env[65788]: DEBUG oslo_concurrency.lockutils [req-428f99fa-076d-4e86-8186-d3889db2584c req-80d86f94-d267-4b71-a1e9-1d8bebabe5db service nova] Releasing lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 983.002408] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquired lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 983.002602] env[65788]: DEBUG nova.network.neutron [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 983.058048] env[65788]: INFO nova.compute.resource_tracker [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating resource usage from migration 761247b6-8ee4-4c67-b8fb-cfd2656b3d8d [ 983.083788] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662828, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.112472] env[65788]: WARNING neutronclient.v2_0.client [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.113398] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.114061] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.403435] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.419318] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8cc71a-f958-474d-a83d-6aa48b139c32 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.433096] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aad63633-8801-4212-9c68-1efc0cd43a1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.446798] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.447899] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.494559] env[65788]: INFO nova.compute.manager [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] instance snapshotting [ 983.501951] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c001383-dc5c-4d61-9756-a599d8f78c8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.505307] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a48de6-41f9-4a4a-9a01-7e08dd960ec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.508991] env[65788]: WARNING neutronclient.v2_0.client [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.509669] env[65788]: WARNING openstack [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.510036] env[65788]: WARNING openstack [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.536104] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4559719b-c947-4a04-bb09-0b51aa4e10bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.541214] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9476ba52-c768-4ba0-a86c-d125815a3c11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.558792] env[65788]: DEBUG nova.compute.provider_tree [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.566931] env[65788]: WARNING neutronclient.v2_0.client [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.567818] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.568342] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.589350] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662828, 'name': CreateSnapshot_Task, 'duration_secs': 0.589707} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.589759] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 983.590763] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364114b6-2816-4c98-b4d8-bbd50d1c7301 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.766166] env[65788]: DEBUG nova.compute.manager [req-191625ac-4b34-40fc-87b3-6ef7aff11251 req-0f844471-9b68-4fb9-9176-2332130a1f2f service nova] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Received event network-vif-deleted-cd2184dd-c826-412e-a259-8e9ebb22c2e5 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 983.774215] env[65788]: DEBUG nova.network.neutron [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 983.787977] env[65788]: WARNING openstack [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.791287] env[65788]: WARNING openstack [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 983.903887] env[65788]: WARNING neutronclient.v2_0.client [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 983.904769] env[65788]: WARNING openstack [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 983.904926] env[65788]: WARNING openstack [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.026711] env[65788]: DEBUG nova.network.neutron [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updating instance_info_cache with network_info: [{"id": "5c238645-7548-466b-983b-2ad80497b15e", "address": "fa:16:3e:38:90:25", "network": {"id": "cab216ab-c10a-47c5-9b6a-c49af536adf5", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-701277193-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22a3db3230244ef5af1227df6f4a6f12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c238645-75", "ovs_interfaceid": "5c238645-7548-466b-983b-2ad80497b15e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 984.079166] env[65788]: DEBUG nova.scheduler.client.report [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 984.084285] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 984.084794] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b8037c22-f625-49ae-ad60-975d02fc1de8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.093583] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 984.093583] env[65788]: value = "task-4662830" [ 984.093583] env[65788]: _type = "Task" [ 984.093583] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.112512] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 984.112915] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662830, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.113551] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e54f7ffa-daf0-4706-8910-fde44a115ecb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.123921] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 984.123921] env[65788]: value = "task-4662831" [ 984.123921] env[65788]: _type = "Task" [ 984.123921] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.134298] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662831, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.278623] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 984.278623] env[65788]: WARNING neutronclient.v2_0.client [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 984.279224] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 984.279553] env[65788]: WARNING openstack [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 984.285244] env[65788]: WARNING neutronclient.v2_0.client [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 984.530237] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Releasing lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 984.586535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.540s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.586671] env[65788]: INFO nova.compute.manager [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Migrating [ 984.598482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.109s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.600419] env[65788]: INFO nova.compute.claims [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.615277] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.621733] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08786130-ab21-4d58-8e65-09655f2333ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.633203] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662830, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.638946] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.638946] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-708698e6-e86e-405b-9653-f635ca6e5473 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.643831] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662831, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.697997] env[65788]: DEBUG nova.compute.manager [req-41cf4197-692d-42dd-b324-6ac9b985f843 req-43cd23d1-4919-499b-92cf-18ebcf4b3354 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-vif-unplugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 984.698838] env[65788]: DEBUG oslo_concurrency.lockutils [req-41cf4197-692d-42dd-b324-6ac9b985f843 req-43cd23d1-4919-499b-92cf-18ebcf4b3354 service nova] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 984.699096] env[65788]: DEBUG oslo_concurrency.lockutils [req-41cf4197-692d-42dd-b324-6ac9b985f843 req-43cd23d1-4919-499b-92cf-18ebcf4b3354 service nova] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 984.699373] env[65788]: DEBUG oslo_concurrency.lockutils [req-41cf4197-692d-42dd-b324-6ac9b985f843 req-43cd23d1-4919-499b-92cf-18ebcf4b3354 service nova] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 984.699808] env[65788]: DEBUG nova.compute.manager [req-41cf4197-692d-42dd-b324-6ac9b985f843 req-43cd23d1-4919-499b-92cf-18ebcf4b3354 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] No waiting events found dispatching network-vif-unplugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 984.699932] env[65788]: WARNING nova.compute.manager [req-41cf4197-692d-42dd-b324-6ac9b985f843 req-43cd23d1-4919-499b-92cf-18ebcf4b3354 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received unexpected event network-vif-unplugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 for instance with vm_state shelved and task_state shelving_offloading. [ 984.717987] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.718397] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.718650] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleting the datastore file [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.718885] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1eef57b8-e893-4876-96fa-d01745845748 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.734239] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 984.734239] env[65788]: value = "task-4662833" [ 984.734239] env[65788]: _type = "Task" [ 984.734239] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.743733] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.037029] env[65788]: DEBUG nova.compute.manager [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 985.037029] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505e5757-d907-477e-8325-4e2ddc7d2a2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.124181] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662830, 'name': CreateSnapshot_Task, 'duration_secs': 0.813945} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.125241] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 985.126286] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.126286] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 985.126372] env[65788]: DEBUG nova.network.neutron [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 985.128526] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f1e7a1-e0d4-4c75-bb1b-1ec9d1545531 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.142278] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662831, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.247221] env[65788]: DEBUG oslo_vmware.api [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184818} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.247571] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.247805] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.247995] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.269903] env[65788]: INFO nova.scheduler.client.report [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted allocations for instance a35ced42-4317-49b4-b4cc-4ed7e2c85c64 [ 985.632685] env[65788]: WARNING neutronclient.v2_0.client [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.633510] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.633890] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.647635] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662831, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.658133] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 985.658986] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9d38ba1d-d875-4098-8b8a-2b3a737ebcc8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.671603] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 985.671603] env[65788]: value = "task-4662834" [ 985.671603] env[65788]: _type = "Task" [ 985.671603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.682423] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662834, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.775650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 985.789143] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.789643] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.966549] env[65788]: WARNING neutronclient.v2_0.client [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 985.967473] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 985.968019] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 985.988284] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbec9d0-c762-47e9-9dd2-cd3cc22c8dea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.998228] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d978e5-719b-405c-a2e5-6190315ac8b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.034286] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b146b40-e494-4e6b-a2ab-4185ee9b8364 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.044114] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca11d30f-c53f-4f9f-a645-239fc8dee330 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.067458] env[65788]: DEBUG nova.compute.provider_tree [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.069777] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b07270a-f8cb-4e94-b387-65ebcfead612 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.086406] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Doing hard reboot of VM {{(pid=65788) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 986.086515] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-fead14d9-94db-46ef-bf8a-61d760a17851 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.096359] env[65788]: DEBUG oslo_vmware.api [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 986.096359] env[65788]: value = "task-4662835" [ 986.096359] env[65788]: _type = "Task" [ 986.096359] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.107518] env[65788]: DEBUG oslo_vmware.api [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662835, 'name': ResetVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.139984] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662831, 'name': CloneVM_Task, 'duration_secs': 1.61375} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.140513] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Created linked-clone VM from snapshot [ 986.142492] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9463d501-3d0c-4385-aa4b-4fedc3df2c66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.152618] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Uploading image 41be19d5-3873-47d3-adea-725252e8e494 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 986.167866] env[65788]: DEBUG nova.network.neutron [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 986.186630] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662834, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.188304] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 986.188304] env[65788]: value = "vm-910366" [ 986.188304] env[65788]: _type = "VirtualMachine" [ 986.188304] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 986.189310] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1a9d7f38-a36d-4e73-b40a-ebae3a70adfd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.198738] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lease: (returnval){ [ 986.198738] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1de14-824b-22e0-20eb-0aedaea145a8" [ 986.198738] env[65788]: _type = "HttpNfcLease" [ 986.198738] env[65788]: } obtained for exporting VM: (result){ [ 986.198738] env[65788]: value = "vm-910366" [ 986.198738] env[65788]: _type = "VirtualMachine" [ 986.198738] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 986.199210] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the lease: (returnval){ [ 986.199210] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1de14-824b-22e0-20eb-0aedaea145a8" [ 986.199210] env[65788]: _type = "HttpNfcLease" [ 986.199210] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 986.210726] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 986.210726] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1de14-824b-22e0-20eb-0aedaea145a8" [ 986.210726] env[65788]: _type = "HttpNfcLease" [ 986.210726] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 986.580969] env[65788]: DEBUG nova.scheduler.client.report [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 986.609935] env[65788]: DEBUG oslo_vmware.api [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662835, 'name': ResetVM_Task, 'duration_secs': 0.120997} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.610604] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Did hard reboot of VM {{(pid=65788) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 986.610802] env[65788]: DEBUG nova.compute.manager [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 986.611645] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44d0720-3227-4f19-8f34-ad831ce992a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.670475] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 986.683626] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662834, 'name': CloneVM_Task} progress is 95%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.708811] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 986.708811] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1de14-824b-22e0-20eb-0aedaea145a8" [ 986.708811] env[65788]: _type = "HttpNfcLease" [ 986.708811] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 986.712277] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 986.712277] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c1de14-824b-22e0-20eb-0aedaea145a8" [ 986.712277] env[65788]: _type = "HttpNfcLease" [ 986.712277] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 986.712277] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0f0265-63e0-4db3-a3b9-c2f7a76c911d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.718578] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c4add4-e18b-961b-2665-79996ba55f88/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 986.718794] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c4add4-e18b-961b-2665-79996ba55f88/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 986.789704] env[65788]: DEBUG nova.compute.manager [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-changed-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 986.789964] env[65788]: DEBUG nova.compute.manager [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Refreshing instance network info cache due to event network-changed-2cd6eb89-f768-4ee6-93a2-386b83c42638. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 986.790204] env[65788]: DEBUG oslo_concurrency.lockutils [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.790351] env[65788]: DEBUG oslo_concurrency.lockutils [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 986.790511] env[65788]: DEBUG nova.network.neutron [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Refreshing network info cache for port 2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 986.829700] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-41e0f965-ba06-430a-877f-8c04199fb0ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.089238] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.089238] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 987.091925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.988s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 987.093941] env[65788]: INFO nova.compute.claims [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.127061] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d53e8be7-45f2-4cae-b51a-e516bf2eee5e tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.874s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 987.194606] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662834, 'name': CloneVM_Task, 'duration_secs': 1.318546} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.195985] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Created linked-clone VM from snapshot [ 987.197909] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736f0c9f-f258-49e4-9ba1-cd7966b25a77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.210497] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Uploading image 15fbd185-1518-47c2-ab66-5548bca182f2 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 987.265466] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 987.265466] env[65788]: value = "vm-910368" [ 987.265466] env[65788]: _type = "VirtualMachine" [ 987.265466] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 987.265466] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8326e226-4a7f-4d63-88be-cf2c86b54476 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.273453] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lease: (returnval){ [ 987.273453] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bf9cb-1b46-eeaa-4ccd-b0a53681c419" [ 987.273453] env[65788]: _type = "HttpNfcLease" [ 987.273453] env[65788]: } obtained for exporting VM: (result){ [ 987.273453] env[65788]: value = "vm-910368" [ 987.273453] env[65788]: _type = "VirtualMachine" [ 987.273453] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 987.273809] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the lease: (returnval){ [ 987.273809] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bf9cb-1b46-eeaa-4ccd-b0a53681c419" [ 987.273809] env[65788]: _type = "HttpNfcLease" [ 987.273809] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 987.283929] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 987.283929] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bf9cb-1b46-eeaa-4ccd-b0a53681c419" [ 987.283929] env[65788]: _type = "HttpNfcLease" [ 987.283929] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 987.295475] env[65788]: WARNING neutronclient.v2_0.client [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.296256] env[65788]: WARNING openstack [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.296902] env[65788]: WARNING openstack [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.501873] env[65788]: WARNING openstack [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.502347] env[65788]: WARNING openstack [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.594536] env[65788]: DEBUG nova.compute.utils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 987.597456] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 987.597747] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 987.598497] env[65788]: WARNING neutronclient.v2_0.client [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.598497] env[65788]: WARNING neutronclient.v2_0.client [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 987.599018] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 987.600782] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 987.611845] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 987.785115] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 987.785115] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bf9cb-1b46-eeaa-4ccd-b0a53681c419" [ 987.785115] env[65788]: _type = "HttpNfcLease" [ 987.785115] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 987.785366] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 987.785366] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bf9cb-1b46-eeaa-4ccd-b0a53681c419" [ 987.785366] env[65788]: _type = "HttpNfcLease" [ 987.785366] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 987.787340] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8e563f-888a-4f83-a5b0-7d75e67d107b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.797674] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5281bb95-e84c-6b85-8430-0353ce0f9f19/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 987.797862] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5281bb95-e84c-6b85-8430-0353ce0f9f19/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 987.912780] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2ccb60b9-fa54-40a2-9adf-e2bcea55f780 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.915371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 988.007359] env[65788]: DEBUG nova.policy [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2886a02e36704e84ac771401e396629f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd2c665795dc4241a50c46de045f2c80', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 988.048214] env[65788]: WARNING neutronclient.v2_0.client [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 988.048425] env[65788]: WARNING openstack [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 988.048630] env[65788]: WARNING openstack [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 988.192456] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fb96a1-84da-4584-b0cc-4cef2b944c9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.218631] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 988.311406] env[65788]: DEBUG nova.network.neutron [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updated VIF entry in instance network info cache for port 2cd6eb89-f768-4ee6-93a2-386b83c42638. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 988.312117] env[65788]: DEBUG nova.network.neutron [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 988.515902] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Successfully created port: e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 988.594903] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4da2da5-4cc5-4983-886a-587f93bd5760 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.610956] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde67eb8-8f99-41ac-b329-53e4dd1c87f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.663269] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 988.668559] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1583ad3c-47d8-4da5-ae15-86b8c482d9b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.679864] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb605de-16b6-4281-8d66-1031013d95ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.697522] env[65788]: DEBUG nova.compute.provider_tree [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 988.739395] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.739395] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34ea7eb8-d76c-40fa-acbe-4790737fbe4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.750698] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 988.750698] env[65788]: value = "task-4662838" [ 988.750698] env[65788]: _type = "Task" [ 988.750698] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.764149] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662838, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.821336] env[65788]: DEBUG oslo_concurrency.lockutils [req-7e8bc9a7-59b6-47d1-a757-8ba2c0c78d7e req-299800e9-6ad8-40d5-8d58-45f4ecd57a33 service nova] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 988.910702] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Successfully created port: 7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 989.123539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "af9120a6-2cea-46da-ba06-6036b9beda78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.125631] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.125631] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "af9120a6-2cea-46da-ba06-6036b9beda78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.125631] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.125631] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 989.127581] env[65788]: INFO nova.compute.manager [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Terminating instance [ 989.235923] env[65788]: ERROR nova.scheduler.client.report [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [req-69054fa8-4e57-4e77-89bc-992d0e4cfb6d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-69054fa8-4e57-4e77-89bc-992d0e4cfb6d"}]} [ 989.261925] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662838, 'name': PowerOffVM_Task, 'duration_secs': 0.353778} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.262357] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.262595] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 989.267455] env[65788]: DEBUG nova.scheduler.client.report [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 989.290967] env[65788]: DEBUG nova.scheduler.client.report [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 989.291242] env[65788]: DEBUG nova.compute.provider_tree [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.306259] env[65788]: DEBUG nova.scheduler.client.report [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 989.329628] env[65788]: DEBUG nova.scheduler.client.report [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 989.631978] env[65788]: DEBUG nova.compute.manager [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 989.632056] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.635929] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df34c20c-d7c4-46d7-a4c2-6718ee2b48c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.647465] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.647827] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-898ffbf0-612d-4830-9116-4737adde9684 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.656841] env[65788]: DEBUG oslo_vmware.api [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 989.656841] env[65788]: value = "task-4662839" [ 989.656841] env[65788]: _type = "Task" [ 989.656841] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.666795] env[65788]: DEBUG oslo_vmware.api [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.668892] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c098f9e4-9986-485d-8967-9f7cc898827d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.677999] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e19d34-0c90-4b53-bb4b-deaadf99c667 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.722824] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405c21c4-2ba2-4094-992b-6c6576a2ebcb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.731995] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a920626b-8f2b-4e20-915e-9c2d3103f645 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.748273] env[65788]: DEBUG nova.compute.provider_tree [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.772435] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 989.772829] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 989.773160] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 989.773481] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 989.773746] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 989.774129] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 989.774440] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.774765] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 989.775010] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 989.775278] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 989.775628] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 989.781717] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69db5d9c-b282-4807-9664-546c39a57cd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.801079] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 989.801079] env[65788]: value = "task-4662840" [ 989.801079] env[65788]: _type = "Task" [ 989.801079] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.812906] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662840, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.168202] env[65788]: DEBUG oslo_vmware.api [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662839, 'name': PowerOffVM_Task, 'duration_secs': 0.233303} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.168482] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.168756] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.169082] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1042e342-8976-474f-a388-1f60c16c4d40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.256981] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.257487] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.257867] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Deleting the datastore file [datastore2] af9120a6-2cea-46da-ba06-6036b9beda78 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.258322] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89b287de-e228-4da5-9093-e58542a0068c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.272671] env[65788]: DEBUG oslo_vmware.api [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 990.272671] env[65788]: value = "task-4662842" [ 990.272671] env[65788]: _type = "Task" [ 990.272671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.285743] env[65788]: DEBUG oslo_vmware.api [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.297557] env[65788]: DEBUG nova.scheduler.client.report [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 990.297557] env[65788]: DEBUG nova.compute.provider_tree [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 114 to 115 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 990.297557] env[65788]: DEBUG nova.compute.provider_tree [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 990.312325] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662840, 'name': ReconfigVM_Task, 'duration_secs': 0.225161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.312703] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 990.568546] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Successfully updated port: e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 990.787183] env[65788]: DEBUG oslo_vmware.api [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247785} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.787596] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.787874] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.788180] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.788468] env[65788]: INFO nova.compute.manager [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Took 1.16 seconds to destroy the instance on the hypervisor. [ 990.788833] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 990.789098] env[65788]: DEBUG nova.compute.manager [-] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 990.789203] env[65788]: DEBUG nova.network.neutron [-] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 990.789470] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.790087] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 990.790421] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 990.802506] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.711s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.803170] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 990.806358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.274s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 990.807974] env[65788]: INFO nova.compute.claims [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.819920] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 990.820197] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 990.820467] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 990.820655] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 990.820918] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 990.821121] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 990.821344] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.821502] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 990.821666] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 990.821831] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 990.822025] env[65788]: DEBUG nova.virt.hardware [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 990.828104] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 990.828493] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-356b54ce-f4d3-4791-84a7-0f5f4725be20 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.844431] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 990.855026] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 990.855026] env[65788]: value = "task-4662843" [ 990.855026] env[65788]: _type = "Task" [ 990.855026] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.867825] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662843, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.312305] env[65788]: DEBUG nova.compute.utils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 991.315131] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 991.315131] env[65788]: DEBUG nova.network.neutron [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 991.315131] env[65788]: WARNING neutronclient.v2_0.client [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.315131] env[65788]: WARNING neutronclient.v2_0.client [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 991.315923] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 991.315923] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 991.367518] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662843, 'name': ReconfigVM_Task, 'duration_secs': 0.225724} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.367869] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 991.369132] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ad1519-31b6-448b-b67e-195b368765cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.378088] env[65788]: DEBUG nova.policy [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c4a00121e214dc9a7b0866c8deee18e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffbe44d63c1d432e97849f15615329e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 991.409270] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.410264] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b06ccc2-e857-43e9-b67b-930cc1b26894 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.435433] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 991.435433] env[65788]: value = "task-4662844" [ 991.435433] env[65788]: _type = "Task" [ 991.435433] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.445841] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662844, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.583876] env[65788]: DEBUG nova.network.neutron [-] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 991.722326] env[65788]: DEBUG nova.network.neutron [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Successfully created port: 9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 991.830301] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 991.950922] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662844, 'name': ReconfigVM_Task, 'duration_secs': 0.485151} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.951275] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.951570] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.086911] env[65788]: INFO nova.compute.manager [-] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Took 1.30 seconds to deallocate network for instance. [ 992.224069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3eded3-84ba-4085-8c32-81b9f2db09fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.235212] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d21e60a-deb7-426c-bf0a-4e451cbba5d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.270705] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ef46a3-0fd0-46c1-854c-9ad866c012de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.280605] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d421b7c8-6ccc-4a75-8359-fbb2fb033b76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.298260] env[65788]: DEBUG nova.compute.provider_tree [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.460047] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ace4cd2-a938-47de-936c-9e3f261df960 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.493640] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e666063-dfcf-4ba8-8ace-37508ca40a82 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.517110] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.594309] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 992.709964] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Successfully updated port: 7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 992.803061] env[65788]: DEBUG nova.scheduler.client.report [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 992.843038] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 993.213691] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.213917] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.214155] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 993.309442] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.310145] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 993.314691] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.952s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.315072] env[65788]: DEBUG nova.objects.instance [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'resources' on Instance uuid 3a06a833-5aaa-4b5d-88b3-8a1d469580af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.317542] env[65788]: DEBUG nova.network.neutron [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Successfully updated port: 9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 993.717296] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.717795] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.764909] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 993.789398] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.789920] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.821110] env[65788]: DEBUG nova.compute.utils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 993.826180] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.826550] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.826765] env[65788]: DEBUG nova.network.neutron [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 993.835076] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 993.835549] env[65788]: DEBUG nova.network.neutron [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 993.836069] env[65788]: WARNING neutronclient.v2_0.client [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.836909] env[65788]: WARNING neutronclient.v2_0.client [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.837096] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.837393] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.885647] env[65788]: WARNING neutronclient.v2_0.client [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 993.886364] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 993.886718] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 993.900543] env[65788]: DEBUG nova.policy [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e095550d08426cb59f11d2295b9a60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '246c0426008f478aa245d006c3c129eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 994.012657] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.013234] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.091470] env[65788]: WARNING neutronclient.v2_0.client [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 994.092169] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.092528] env[65788]: WARNING openstack [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.143466] env[65788]: WARNING neutronclient.v2_0.client [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 994.186102] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87aeee2-0ae5-4c63-b0f9-9b3d65cb3c96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.198486] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262893cf-331b-40dd-a15a-c3ae84dc67b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.203780] env[65788]: DEBUG nova.network.neutron [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Port e98f1b15-0a05-4be4-b950-dd7e6749eb8f binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 994.237427] env[65788]: DEBUG nova.network.neutron [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Successfully created port: 20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 994.241363] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee5a036-b35d-474e-936b-64ff3ee9a238 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.247302] env[65788]: DEBUG nova.network.neutron [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updating instance_info_cache with network_info: [{"id": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "address": "fa:16:3e:e0:3f:54", "network": {"id": "3845b505-1663-4cb0-88cc-f080dee79756", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2071292552", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5413129-8b", "ovs_interfaceid": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "address": "fa:16:3e:5f:23:09", "network": {"id": "e3030d2a-0cd6-4c50-82ae-6481cd869a29", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1516022772", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b27f93c-6a", "ovs_interfaceid": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 994.255563] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f266e2c-6c1f-4706-9060-2a123facd64b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.274189] env[65788]: DEBUG nova.compute.provider_tree [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.336205] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 994.348520] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.349129] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.399810] env[65788]: DEBUG nova.network.neutron [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 994.421235] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.421666] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.490494] env[65788]: WARNING neutronclient.v2_0.client [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 994.491299] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 994.491663] env[65788]: WARNING openstack [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 994.580326] env[65788]: DEBUG nova.network.neutron [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Updating instance_info_cache with network_info: [{"id": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "address": "fa:16:3e:24:47:96", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f372d9b-f1", "ovs_interfaceid": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 994.680115] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 994.680455] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 994.680648] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 994.680946] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 994.681613] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 994.681613] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 994.681745] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.681877] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 994.682024] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 994.682473] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 994.682473] env[65788]: DEBUG nova.virt.hardware [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 994.684841] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb980d57-2e0f-40ff-af42-e910f8954102 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.694338] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f54fc0d-d5b5-4b4a-94be-3773099cf651 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.750217] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Releasing lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.750681] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Instance network_info: |[{"id": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "address": "fa:16:3e:e0:3f:54", "network": {"id": "3845b505-1663-4cb0-88cc-f080dee79756", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2071292552", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5413129-8b", "ovs_interfaceid": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "address": "fa:16:3e:5f:23:09", "network": {"id": "e3030d2a-0cd6-4c50-82ae-6481cd869a29", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1516022772", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b27f93c-6a", "ovs_interfaceid": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 994.751281] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:3f:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '674802e7-b847-4bef-a7a8-f90ac7a3a0a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5413129-8bf0-444f-9791-900f8e5c2f6e', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:23:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89ed4797-90ad-44cd-bbcb-e90b2a8400f3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b27f93c-6a27-456b-b3aa-04d0cc66c075', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.761288] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 994.761909] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.761909] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-21771c35-9c4d-4da9-a631-2856d6158666 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.784726] env[65788]: DEBUG nova.scheduler.client.report [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 994.795324] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.795324] env[65788]: value = "task-4662845" [ 994.795324] env[65788]: _type = "Task" [ 994.795324] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.808892] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662845, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.013832] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 995.014174] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 995.014405] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 995.014743] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 995.014800] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 995.014946] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 995.015258] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.015472] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 995.015653] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 995.015837] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 995.016244] env[65788]: DEBUG nova.virt.hardware [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 995.018940] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1198b850-fc32-476a-b8d9-a35426bf1a95 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.030972] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c4add4-e18b-961b-2665-79996ba55f88/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 995.032495] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c79360-7a80-4459-afa4-cecd5c12419e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.041255] env[65788]: DEBUG nova.compute.manager [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Received event network-changed-5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 995.041326] env[65788]: DEBUG nova.compute.manager [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Refreshing instance network info cache due to event network-changed-5c238645-7548-466b-983b-2ad80497b15e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 995.042196] env[65788]: DEBUG oslo_concurrency.lockutils [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] Acquiring lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.042196] env[65788]: DEBUG oslo_concurrency.lockutils [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] Acquired lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.043043] env[65788]: DEBUG nova.network.neutron [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Refreshing network info cache for port 5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 995.044828] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f994fad-9bb5-43b8-9f6c-84cf33105090 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.052489] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c4add4-e18b-961b-2665-79996ba55f88/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 995.052489] env[65788]: ERROR oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c4add4-e18b-961b-2665-79996ba55f88/disk-0.vmdk due to incomplete transfer. [ 995.053207] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2fa6170d-2f20-4635-be11-b1160e00e034 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.073317] env[65788]: DEBUG oslo_vmware.rw_handles [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c4add4-e18b-961b-2665-79996ba55f88/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 995.073534] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Uploaded image 41be19d5-3873-47d3-adea-725252e8e494 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 995.076170] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 995.076475] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-50dedfee-379f-4f55-ae65-df7c0eae2542 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.083179] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.083570] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Instance network_info: |[{"id": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "address": "fa:16:3e:24:47:96", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f372d9b-f1", "ovs_interfaceid": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 995.086066] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:47:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f372d9b-f11d-40fc-b795-bae66f4a1800', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.093111] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 995.093471] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 995.093471] env[65788]: value = "task-4662846" [ 995.093471] env[65788]: _type = "Task" [ 995.093471] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.094230] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 995.094645] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e81ecf1c-b9ea-49a9-a401-02cd79f28789 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.121357] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662846, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.122843] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.122843] env[65788]: value = "task-4662847" [ 995.122843] env[65788]: _type = "Task" [ 995.122843] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.133092] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662847, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.232711] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.233090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.233296] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.290423] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.977s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.294963] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.581s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.294963] env[65788]: DEBUG nova.objects.instance [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lazy-loading 'resources' on Instance uuid b220df77-cd3a-4aeb-87ad-68f56995ec3b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.311392] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662845, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.314963] env[65788]: INFO nova.scheduler.client.report [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleted allocations for instance 3a06a833-5aaa-4b5d-88b3-8a1d469580af [ 995.346762] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 995.381934] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 995.382297] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 995.383088] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 995.383088] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 995.383088] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 995.383088] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 995.383535] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.383773] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 995.384039] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 995.384237] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 995.384437] env[65788]: DEBUG nova.virt.hardware [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 995.385908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6201dc4c-cb53-4f66-8d3d-8730d9abb52a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.396777] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04044fa-e77c-43f6-8a6a-c7a858c04f05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.529098] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.529224] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.529419] env[65788]: INFO nova.compute.manager [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Shelving [ 995.550187] env[65788]: WARNING neutronclient.v2_0.client [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 995.550923] env[65788]: WARNING openstack [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 995.551280] env[65788]: WARNING openstack [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 995.605358] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662846, 'name': Destroy_Task, 'duration_secs': 0.493554} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.605724] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Destroyed the VM [ 995.606034] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 995.606362] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0a250ff5-60bb-4bfc-a451-1fe223f31bf0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.614240] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 995.614240] env[65788]: value = "task-4662848" [ 995.614240] env[65788]: _type = "Task" [ 995.614240] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.623542] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662848, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.633645] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662847, 'name': CreateVM_Task, 'duration_secs': 0.457548} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.633983] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.634588] env[65788]: WARNING neutronclient.v2_0.client [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 995.635044] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.635282] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 995.635673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 995.636028] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5eadc6c-8a87-49c6-a904-bb4098577233 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.641823] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 995.641823] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528ebe09-8e54-f662-138a-bb05e80c965e" [ 995.641823] env[65788]: _type = "Task" [ 995.641823] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.651133] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528ebe09-8e54-f662-138a-bb05e80c965e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.669101] env[65788]: DEBUG nova.network.neutron [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 995.709539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "6e78042d-ba10-4629-816f-3a13b2e22d4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.710489] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.754241] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5281bb95-e84c-6b85-8430-0353ce0f9f19/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 995.755612] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e50ed29-a759-49ff-937a-67226047e15c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.764620] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5281bb95-e84c-6b85-8430-0353ce0f9f19/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 995.764825] env[65788]: ERROR oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5281bb95-e84c-6b85-8430-0353ce0f9f19/disk-0.vmdk due to incomplete transfer. [ 995.766296] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-32332b85-db3f-45ae-9695-d909ce35d0b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.774999] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "51705add-4d88-40bb-b236-2486650c77f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.775249] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "51705add-4d88-40bb-b236-2486650c77f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.783784] env[65788]: DEBUG oslo_vmware.rw_handles [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5281bb95-e84c-6b85-8430-0353ce0f9f19/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 995.784580] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Uploaded image 15fbd185-1518-47c2-ab66-5548bca182f2 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 995.785896] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 995.787155] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4d17b2f4-f63a-4ad2-9c38-41d03182998e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.811336] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 995.811336] env[65788]: value = "task-4662849" [ 995.811336] env[65788]: _type = "Task" [ 995.811336] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.833462] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662845, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.834289] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3cc468c3-6929-47fd-be53-fdc3109f35fb tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "3a06a833-5aaa-4b5d-88b3-8a1d469580af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.187s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.838730] env[65788]: DEBUG nova.network.neutron [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Successfully updated port: 20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 995.847182] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662849, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.880188] env[65788]: DEBUG nova.network.neutron [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 995.884681] env[65788]: DEBUG nova.compute.manager [req-6b3c8c7c-5f15-4d46-8563-776deb1036e5 req-e27eaccb-b81e-4185-88dd-50d15654c518 service nova] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Received event network-vif-deleted-5c238645-7548-466b-983b-2ad80497b15e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 996.127499] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662848, 'name': RemoveSnapshot_Task, 'duration_secs': 0.458137} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.127821] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 996.128250] env[65788]: DEBUG nova.compute.manager [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 996.129624] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f269482-de1c-4f2d-aeb7-24cee743cf5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.154917] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528ebe09-8e54-f662-138a-bb05e80c965e, 'name': SearchDatastore_Task, 'duration_secs': 0.01442} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.155280] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.155501] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.155735] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.155894] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.156104] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 996.156495] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d763ab4b-8e93-4821-be11-defafd39b164 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.167440] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 996.167712] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 996.168561] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47c8229b-4398-44b3-b074-aba1382fb4f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.175103] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 996.175103] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b64637-da44-cc68-7db7-895c38414871" [ 996.175103] env[65788]: _type = "Task" [ 996.175103] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.183910] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b64637-da44-cc68-7db7-895c38414871, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.201708] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde3a5d7-bc03-4482-961a-5b9daaf38b5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.210944] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1825a92f-7a98-4120-9cbb-fe1b5fcbc3f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.214781] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 996.252945] env[65788]: WARNING neutronclient.v2_0.client [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.256907] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e30a411-88dd-43c6-ab04-2feec13abc5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.267675] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9516eb61-1e13-431e-9536-44ac31418b0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.285135] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 996.288624] env[65788]: DEBUG nova.compute.provider_tree [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.313693] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662845, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.324373] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662849, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.340524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "refresh_cache-2bcfbe11-51a2-49b4-b482-02ca332d8c38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.340978] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "refresh_cache-2bcfbe11-51a2-49b4-b482-02ca332d8c38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.340978] env[65788]: DEBUG nova.network.neutron [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 996.385493] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.385493] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.385493] env[65788]: DEBUG nova.network.neutron [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 996.386537] env[65788]: DEBUG oslo_concurrency.lockutils [req-f0581e13-af50-46f6-9ddd-8f9fbbd38a6c req-c77f94da-b717-446b-b84e-adf3f28c15ec service nova] Releasing lock "refresh_cache-af9120a6-2cea-46da-ba06-6036b9beda78" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.497147] env[65788]: DEBUG oslo_concurrency.lockutils [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.497379] env[65788]: DEBUG oslo_concurrency.lockutils [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.547131] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.547616] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6af79802-2b7e-464f-84b5-89399a55d953 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.556222] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 996.556222] env[65788]: value = "task-4662851" [ 996.556222] env[65788]: _type = "Task" [ 996.556222] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.568917] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.651686] env[65788]: INFO nova.compute.manager [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Shelve offloading [ 996.686249] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b64637-da44-cc68-7db7-895c38414871, 'name': SearchDatastore_Task, 'duration_secs': 0.011069} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.687195] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b2500b0-7976-46ef-934b-dc58e66c01dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.693929] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 996.693929] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5284acd2-15bc-156c-0b5a-b53a0f8e4baa" [ 996.693929] env[65788]: _type = "Task" [ 996.693929] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.703356] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5284acd2-15bc-156c-0b5a-b53a0f8e4baa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.741787] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.797439] env[65788]: DEBUG nova.scheduler.client.report [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 996.813948] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662845, 'name': CreateVM_Task, 'duration_secs': 1.521799} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.815120] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.818826] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 996.819669] env[65788]: WARNING neutronclient.v2_0.client [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.820053] env[65788]: WARNING neutronclient.v2_0.client [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.820381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.820535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.820862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 996.821513] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4c4253b-4da0-421a-8bd3-956f6b33d628 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.829087] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662849, 'name': Destroy_Task, 'duration_secs': 0.792411} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.830438] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Destroyed the VM [ 996.830695] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 996.831069] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 996.831069] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52142916-d7c7-f186-6b3a-8c9f02ddf21e" [ 996.831069] env[65788]: _type = "Task" [ 996.831069] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.831278] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fb1f85bb-81ec-44c3-b35e-02528c322d29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.841330] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52142916-d7c7-f186-6b3a-8c9f02ddf21e, 'name': SearchDatastore_Task, 'duration_secs': 0.009459} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.842469] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 996.842689] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 996.842988] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.843291] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 996.843291] env[65788]: value = "task-4662852" [ 996.843291] env[65788]: _type = "Task" [ 996.843291] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.844037] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.844449] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.860257] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662852, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.880998] env[65788]: DEBUG nova.network.neutron [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 996.887902] env[65788]: WARNING neutronclient.v2_0.client [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.888568] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.889496] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.899951] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.900524] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 996.979084] env[65788]: WARNING neutronclient.v2_0.client [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 996.979751] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 996.980213] env[65788]: WARNING openstack [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.000913] env[65788]: INFO nova.compute.manager [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Detaching volume 1294ba3f-809b-4d9c-be6d-9441e471e9ea [ 997.015028] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.015028] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.040680] env[65788]: INFO nova.virt.block_device [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Attempting to driver detach volume 1294ba3f-809b-4d9c-be6d-9441e471e9ea from mountpoint /dev/sdb [ 997.040985] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 997.041120] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910332', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'name': 'volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd8103991-fd1f-4a57-81ce-1a47dc4defe0', 'attached_at': '', 'detached_at': '', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'serial': '1294ba3f-809b-4d9c-be6d-9441e471e9ea'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 997.042102] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8d58be-f8c3-44bc-ad75-718285dda98a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.078056] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d823c8a-2483-49e9-9452-f32eb2bd0169 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.086046] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662851, 'name': PowerOffVM_Task, 'duration_secs': 0.204624} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.088525] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.089588] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7129e34e-767b-46f6-8be8-66ee652c7513 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.093414] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0cc2bc-a39e-418a-ad52-63ccabdb447c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.100386] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received event network-vif-plugged-e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 997.100566] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquiring lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.100801] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.101099] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.101300] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] No waiting events found dispatching network-vif-plugged-e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 997.101462] env[65788]: WARNING nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received unexpected event network-vif-plugged-e5413129-8bf0-444f-9791-900f8e5c2f6e for instance with vm_state building and task_state spawning. [ 997.101623] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received event network-changed-e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 997.101944] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Refreshing instance network info cache due to event network-changed-e5413129-8bf0-444f-9791-900f8e5c2f6e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 997.102025] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquiring lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.102368] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquired lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.102368] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Refreshing network info cache for port e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 997.106130] env[65788]: DEBUG nova.network.neutron [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Updating instance_info_cache with network_info: [{"id": "20da233d-2947-43a6-88b1-229b364fbd48", "address": "fa:16:3e:52:cf:5b", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20da233d-29", "ovs_interfaceid": "20da233d-2947-43a6-88b1-229b364fbd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 997.132565] env[65788]: WARNING neutronclient.v2_0.client [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 997.133642] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.134057] env[65788]: WARNING openstack [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.158948] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56b6be1-6056-4a3a-8fe8-ed81545b0b6d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.162449] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.163261] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cc80768-59c1-4981-bff8-26bb68557b0d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.165442] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0e607f-739c-492b-bafa-b4cba64b0cb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.190265] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 997.190265] env[65788]: value = "task-4662853" [ 997.190265] env[65788]: _type = "Task" [ 997.190265] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.191170] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The volume has not been displaced from its original location: [datastore1] volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea/volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 997.196797] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 997.197649] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3860edfc-ffe6-4227-8337-329ce12a21d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.232365] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5284acd2-15bc-156c-0b5a-b53a0f8e4baa, 'name': SearchDatastore_Task, 'duration_secs': 0.013957} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.236690] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.237037] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 1a701758-a056-4948-9069-2a7168f2cc8c/1a701758-a056-4948-9069-2a7168f2cc8c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.237543] env[65788]: DEBUG oslo_vmware.api [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 997.237543] env[65788]: value = "task-4662854" [ 997.237543] env[65788]: _type = "Task" [ 997.237543] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.237819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.238085] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.238686] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8d71513-89cf-49d2-9e4f-dbac45793132 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.241380] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7a4a25d-f5db-4105-a693-24f2e2b57b3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.254155] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 997.254155] env[65788]: value = "task-4662855" [ 997.254155] env[65788]: _type = "Task" [ 997.254155] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.259506] env[65788]: DEBUG oslo_vmware.api [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.265699] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.265992] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 997.266774] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d313cafd-ba5c-4d28-8432-1f9c3a204ad0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.271745] env[65788]: DEBUG nova.network.neutron [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 997.279467] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.281244] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 997.281244] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a9fe70-2768-d44b-2cc0-bfca680a7a24" [ 997.281244] env[65788]: _type = "Task" [ 997.281244] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.294619] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a9fe70-2768-d44b-2cc0-bfca680a7a24, 'name': SearchDatastore_Task, 'duration_secs': 0.011964} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.295749] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02b73032-0025-4e9a-8e8d-a5b120b260ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.304036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.010s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.306604] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 997.306604] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d4d45f-7d7e-c45e-4747-75867efa9e66" [ 997.306604] env[65788]: _type = "Task" [ 997.306604] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.307949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.858s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.311439] env[65788]: INFO nova.compute.claims [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.324956] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d4d45f-7d7e-c45e-4747-75867efa9e66, 'name': SearchDatastore_Task, 'duration_secs': 0.011186} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.325635] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.326382] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf/4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 997.326795] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20b2b87b-ae61-479b-9953-93c0cc6cfa1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.337049] env[65788]: INFO nova.scheduler.client.report [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Deleted allocations for instance b220df77-cd3a-4aeb-87ad-68f56995ec3b [ 997.340867] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 997.340867] env[65788]: value = "task-4662856" [ 997.340867] env[65788]: _type = "Task" [ 997.340867] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.356362] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662856, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.372607] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662852, 'name': RemoveSnapshot_Task} progress is 29%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.611030] env[65788]: WARNING neutronclient.v2_0.client [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 997.611030] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.611030] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.618585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "refresh_cache-2bcfbe11-51a2-49b4-b482-02ca332d8c38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.618967] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Instance network_info: |[{"id": "20da233d-2947-43a6-88b1-229b364fbd48", "address": "fa:16:3e:52:cf:5b", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20da233d-29", "ovs_interfaceid": "20da233d-2947-43a6-88b1-229b364fbd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 997.619705] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:cf:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a58387dd-f438-4913-af6a-fafb734cd881', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20da233d-2947-43a6-88b1-229b364fbd48', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.628323] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 997.628668] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 997.629051] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1369d86e-6fdb-4a2a-980e-37b42f32595c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.654548] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.654548] env[65788]: value = "task-4662857" [ 997.654548] env[65788]: _type = "Task" [ 997.654548] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.668267] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662857, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.698939] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 997.699393] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-aaa26daf-9e46-4866-91e1-f6091504b994 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.713669] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 997.713669] env[65788]: value = "task-4662858" [ 997.713669] env[65788]: _type = "Task" [ 997.713669] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.723968] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 997.724348] env[65788]: DEBUG nova.compute.manager [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 997.725860] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3b6c4f-c13c-4e7b-93d0-09d3e8151f05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.736398] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.736643] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.736874] env[65788]: DEBUG nova.network.neutron [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 997.742457] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662858, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.763977] env[65788]: DEBUG oslo_vmware.api [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662854, 'name': ReconfigVM_Task, 'duration_secs': 0.278587} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.766488] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 997.775218] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.775218] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 997.781856] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b850129-44f7-46d6-84d1-13c9b02ce3f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.793219] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.807234] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662855, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.808784] env[65788]: DEBUG oslo_vmware.api [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 997.808784] env[65788]: value = "task-4662859" [ 997.808784] env[65788]: _type = "Task" [ 997.808784] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.823215] env[65788]: DEBUG oslo_vmware.api [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662859, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.858925] env[65788]: DEBUG oslo_concurrency.lockutils [None req-73380b3a-cb0c-4d5d-b778-ba8f35245af2 tempest-ImagesTestJSON-433702528 tempest-ImagesTestJSON-433702528-project-member] Lock "b220df77-cd3a-4aeb-87ad-68f56995ec3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.669s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.868251] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662856, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.878930] env[65788]: DEBUG oslo_vmware.api [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662852, 'name': RemoveSnapshot_Task, 'duration_secs': 1.03287} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.879383] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 997.879727] env[65788]: INFO nova.compute.manager [None req-f5c0d156-087e-49e7-8349-5ab7ac430c32 tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Took 14.38 seconds to snapshot the instance on the hypervisor. [ 997.907951] env[65788]: WARNING neutronclient.v2_0.client [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 997.908733] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 997.909106] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.008664] env[65788]: DEBUG nova.compute.manager [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Received event network-vif-plugged-9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 998.010064] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Acquiring lock "1a701758-a056-4948-9069-2a7168f2cc8c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.010064] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Lock "1a701758-a056-4948-9069-2a7168f2cc8c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.010064] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Lock "1a701758-a056-4948-9069-2a7168f2cc8c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.010064] env[65788]: DEBUG nova.compute.manager [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] No waiting events found dispatching network-vif-plugged-9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 998.010064] env[65788]: WARNING nova.compute.manager [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Received unexpected event network-vif-plugged-9f372d9b-f11d-40fc-b795-bae66f4a1800 for instance with vm_state building and task_state spawning. [ 998.010366] env[65788]: DEBUG nova.compute.manager [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Received event network-changed-9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 998.010366] env[65788]: DEBUG nova.compute.manager [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Refreshing instance network info cache due to event network-changed-9f372d9b-f11d-40fc-b795-bae66f4a1800. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 998.010479] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Acquiring lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.010697] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Acquired lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.010918] env[65788]: DEBUG nova.network.neutron [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Refreshing network info cache for port 9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 998.048054] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updated VIF entry in instance network info cache for port e5413129-8bf0-444f-9791-900f8e5c2f6e. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 998.048440] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updating instance_info_cache with network_info: [{"id": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "address": "fa:16:3e:e0:3f:54", "network": {"id": "3845b505-1663-4cb0-88cc-f080dee79756", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2071292552", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5413129-8b", "ovs_interfaceid": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "address": "fa:16:3e:5f:23:09", "network": {"id": "e3030d2a-0cd6-4c50-82ae-6481cd869a29", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1516022772", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b27f93c-6a", "ovs_interfaceid": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 998.166906] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662857, 'name': CreateVM_Task, 'duration_secs': 0.445168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.167106] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 998.167692] env[65788]: WARNING neutronclient.v2_0.client [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.167982] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.168154] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.168504] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 998.168799] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edeae09f-cee6-4b89-a37e-59a7d29cbf68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.175052] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 998.175052] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a3e541-5d0f-b7aa-2280-780dc304209e" [ 998.175052] env[65788]: _type = "Task" [ 998.175052] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.184207] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a3e541-5d0f-b7aa-2280-780dc304209e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.230049] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662858, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.248889] env[65788]: WARNING neutronclient.v2_0.client [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.249630] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.249990] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.273100] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579999} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.273274] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 1a701758-a056-4948-9069-2a7168f2cc8c/1a701758-a056-4948-9069-2a7168f2cc8c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.273468] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.273723] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-959d12b4-26f2-48a3-8584-0ad3a805610c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.281851] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 998.281851] env[65788]: value = "task-4662860" [ 998.281851] env[65788]: _type = "Task" [ 998.281851] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.291770] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662860, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.304451] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381c653a-297c-4057-bae8-50f339366d91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.316094] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423c7122-3df7-4a3c-ba65-779bbc81149e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.322426] env[65788]: DEBUG oslo_vmware.api [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662859, 'name': ReconfigVM_Task, 'duration_secs': 0.20223} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.326142] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910332', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'name': 'volume-1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd8103991-fd1f-4a57-81ce-1a47dc4defe0', 'attached_at': '', 'detached_at': '', 'volume_id': '1294ba3f-809b-4d9c-be6d-9441e471e9ea', 'serial': '1294ba3f-809b-4d9c-be6d-9441e471e9ea'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 998.363226] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662856, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.868626} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.363554] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf/4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 998.363952] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.364337] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a315d099-3e45-4e9d-895a-9464b4af0a56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.372961] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 998.372961] env[65788]: value = "task-4662861" [ 998.372961] env[65788]: _type = "Task" [ 998.372961] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.378669] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.379375] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.396058] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662861, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.457415] env[65788]: WARNING neutronclient.v2_0.client [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.458387] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.458648] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.513709] env[65788]: WARNING neutronclient.v2_0.client [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.514590] env[65788]: WARNING openstack [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.515012] env[65788]: WARNING openstack [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.552187] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Releasing lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.552493] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received event network-vif-plugged-7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 998.552745] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquiring lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 998.553044] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 998.553249] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.553401] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] No waiting events found dispatching network-vif-plugged-7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 998.553564] env[65788]: WARNING nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received unexpected event network-vif-plugged-7b27f93c-6a27-456b-b3aa-04d0cc66c075 for instance with vm_state building and task_state spawning. [ 998.553793] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received event network-changed-7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 998.554760] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Refreshing instance network info cache due to event network-changed-7b27f93c-6a27-456b-b3aa-04d0cc66c075. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 998.555131] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquiring lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.556900] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquired lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.556900] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Refreshing network info cache for port 7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 998.566211] env[65788]: DEBUG nova.network.neutron [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updating instance_info_cache with network_info: [{"id": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "address": "fa:16:3e:bd:34:66", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fb26674-54", "ovs_interfaceid": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 998.690141] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a3e541-5d0f-b7aa-2280-780dc304209e, 'name': SearchDatastore_Task, 'duration_secs': 0.023249} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.693331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.693669] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.693946] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.694170] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 998.694424] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.695199] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8e2723d-d5c3-4600-b579-d88a1a75bcc0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.716045] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.716476] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 998.717441] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0cf047f-79c8-4867-90ea-55f864be0817 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.733331] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 998.733331] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bebf6-2fc9-6c45-4d65-d00e17cf111b" [ 998.733331] env[65788]: _type = "Task" [ 998.733331] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.738489] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662858, 'name': CreateSnapshot_Task, 'duration_secs': 1.013555} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.742505] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 998.744487] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc7a12f-a572-4832-a1c9-0403776fe3f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.749186] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0458928e-3a8d-4b0f-9b48-6ad3463335de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.753716] env[65788]: WARNING openstack [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.754160] env[65788]: WARNING openstack [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.777224] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525bebf6-2fc9-6c45-4d65-d00e17cf111b, 'name': SearchDatastore_Task, 'duration_secs': 0.013393} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.778592] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305d469b-6027-4792-b6c9-cb39d28d416b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.783277] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee5b71ad-b90e-4b21-b4da-dbeabdab9001 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.821461] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46cc26b8-ad7a-4d7f-be0a-d98b7adc3828 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.835185] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 998.835185] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5206d8fd-7a8b-6f02-8dd5-7d875a7ba8a6" [ 998.835185] env[65788]: _type = "Task" [ 998.835185] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.835369] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662860, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098954} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.836318] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.840679] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d03ea5f-2d09-42fb-83a1-f551dfe06623 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.847267] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6434ac58-0d92-4c24-9190-6d51a08b75a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.875755] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 1a701758-a056-4948-9069-2a7168f2cc8c/1a701758-a056-4948-9069-2a7168f2cc8c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.878101] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5206d8fd-7a8b-6f02-8dd5-7d875a7ba8a6, 'name': SearchDatastore_Task, 'duration_secs': 0.009886} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.878101] env[65788]: WARNING neutronclient.v2_0.client [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 998.878467] env[65788]: WARNING openstack [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 998.878842] env[65788]: WARNING openstack [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 998.889698] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f66e7ab-30b6-479c-8d37-dbc661c78826 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.912986] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.913420] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2bcfbe11-51a2-49b4-b482-02ca332d8c38/2bcfbe11-51a2-49b4-b482-02ca332d8c38.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.924504] env[65788]: DEBUG nova.objects.instance [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.933461] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d6dcc65-7c22-4246-8106-75ac4ac1862d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.936458] env[65788]: DEBUG nova.compute.provider_tree [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 998.951938] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662861, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072471} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.951938] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 998.951938] env[65788]: value = "task-4662863" [ 998.951938] env[65788]: _type = "Task" [ 998.951938] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.951938] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 998.951938] env[65788]: value = "task-4662862" [ 998.951938] env[65788]: _type = "Task" [ 998.951938] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.952338] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.953138] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9c2da7-7ea0-4ff1-8945-fa779e06ec4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.988995] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662862, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.989335] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.001026] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf/4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.011098] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8889ab3e-8b80-458e-b658-00801988029f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.038453] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 999.038453] env[65788]: value = "task-4662864" [ 999.038453] env[65788]: _type = "Task" [ 999.038453] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.051356] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662864, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.060426] env[65788]: WARNING neutronclient.v2_0.client [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.060561] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.061047] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.075318] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 999.075857] env[65788]: WARNING neutronclient.v2_0.client [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.076922] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.077194] env[65788]: WARNING openstack [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.083346] env[65788]: WARNING neutronclient.v2_0.client [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.101539] env[65788]: DEBUG nova.network.neutron [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Updated VIF entry in instance network info cache for port 9f372d9b-f11d-40fc-b795-bae66f4a1800. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 999.101539] env[65788]: DEBUG nova.network.neutron [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Updating instance_info_cache with network_info: [{"id": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "address": "fa:16:3e:24:47:96", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f372d9b-f1", "ovs_interfaceid": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 999.285944] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 999.290504] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-289bb76b-09ee-4ea4-8aaf-becee7601d71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.305035] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 999.305035] env[65788]: value = "task-4662865" [ 999.305035] env[65788]: _type = "Task" [ 999.305035] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.314429] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662865, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.437250] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.437250] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.470284] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662862, 'name': ReconfigVM_Task, 'duration_secs': 0.351694} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.471720] env[65788]: ERROR nova.scheduler.client.report [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [req-552368ba-3060-4462-854d-37b0174358be] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-552368ba-3060-4462-854d-37b0174358be"}]} [ 999.477305] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 1a701758-a056-4948-9069-2a7168f2cc8c/1a701758-a056-4948-9069-2a7168f2cc8c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.480980] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662863, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.482181] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87d13b7b-f95f-4303-ba70-41fdb0df830c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.492786] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 999.492786] env[65788]: value = "task-4662866" [ 999.492786] env[65788]: _type = "Task" [ 999.492786] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.499232] env[65788]: DEBUG nova.scheduler.client.report [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 999.505071] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662866, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.515591] env[65788]: DEBUG nova.scheduler.client.report [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 999.515823] env[65788]: DEBUG nova.compute.provider_tree [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 999.535063] env[65788]: DEBUG nova.scheduler.client.report [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 999.552726] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662864, 'name': ReconfigVM_Task, 'duration_secs': 0.363744} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.552726] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf/4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.554205] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2799388-4e59-40ab-9aad-b76704255674 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.558252] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436b14c2-f8a5-45c0-aae9-0af3a0b9fa0b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.563148] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 999.563148] env[65788]: value = "task-4662867" [ 999.563148] env[65788]: _type = "Task" [ 999.563148] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.567818] env[65788]: DEBUG nova.scheduler.client.report [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 999.588221] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dd0f3a-fef2-4b80-94fe-39e96e6b6a7c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.595286] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662867, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.602269] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 999.607211] env[65788]: DEBUG oslo_concurrency.lockutils [req-ffe21608-cff4-4e29-a039-8c91ec60e20a req-15c798fe-73aa-4526-a907-cdd72966d944 service nova] Releasing lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 999.683661] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.685609] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29cd117-623c-47b0-8be3-9b144430bcc8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.697170] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.700550] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24e824ae-6d5d-4030-a456-f5dce53067d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.788792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.790373] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.790805] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleting the datastore file [datastore2] 29c0ab17-5df2-4f6c-bdea-3a4a6c857066 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.792114] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef372a07-4d7f-44e5-b4c0-d741c0a34b62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.800908] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 999.800908] env[65788]: value = "task-4662869" [ 999.800908] env[65788]: _type = "Task" [ 999.800908] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.819935] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662869, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.821245] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662865, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.823208] env[65788]: WARNING neutronclient.v2_0.client [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 999.824267] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 999.824360] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 999.953084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-01cc9d57-2824-4954-8bcf-ada522b862ea tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.456s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.969722] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662863, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588975} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.973035] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2bcfbe11-51a2-49b4-b482-02ca332d8c38/2bcfbe11-51a2-49b4-b482-02ca332d8c38.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.973035] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.973035] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c396bf55-3b4c-4151-9520-c1dea87e4892 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.985441] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 999.985441] env[65788]: value = "task-4662870" [ 999.985441] env[65788]: _type = "Task" [ 999.985441] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.005380] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.005380] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e21cc67-8119-4b6d-84bc-e49606a4839b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.013621] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662866, 'name': Rename_Task, 'duration_secs': 0.174169} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.014521] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.014827] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29f8421f-83ff-4874-ae78-f645675fb42f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.020810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4893f1f4-1e96-4333-845f-823304d2a571 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.026534] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1000.026534] env[65788]: value = "task-4662871" [ 1000.026534] env[65788]: _type = "Task" [ 1000.026534] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.072856] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updated VIF entry in instance network info cache for port 7b27f93c-6a27-456b-b3aa-04d0cc66c075. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1000.073520] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updating instance_info_cache with network_info: [{"id": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "address": "fa:16:3e:e0:3f:54", "network": {"id": "3845b505-1663-4cb0-88cc-f080dee79756", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2071292552", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5413129-8b", "ovs_interfaceid": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "address": "fa:16:3e:5f:23:09", "network": {"id": "e3030d2a-0cd6-4c50-82ae-6481cd869a29", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1516022772", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b27f93c-6a", "ovs_interfaceid": "7b27f93c-6a27-456b-b3aa-04d0cc66c075", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1000.085707] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7b9f29-8ed2-45e8-85e5-ab82f6d93cf0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.091110] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662871, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.100648] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662867, 'name': Rename_Task, 'duration_secs': 0.181812} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.102670] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c545109-b8ee-4f73-8013-c72d5ed10e32 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.106841] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.107920] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9b57f24-0cbf-4d1c-995f-c1fc834e1b92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.113670] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1000.113670] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d739dc82-bf48-450f-9a52-f90d76000dba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.123663] env[65788]: DEBUG nova.compute.provider_tree [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.126522] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 1000.126522] env[65788]: value = "task-4662872" [ 1000.126522] env[65788]: _type = "Task" [ 1000.126522] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.133872] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1000.133872] env[65788]: value = "task-4662873" [ 1000.133872] env[65788]: _type = "Task" [ 1000.133872] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.136066] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.147728] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.317899] env[65788]: DEBUG oslo_vmware.api [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662869, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.434024} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.322072] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.322468] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.322836] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.325798] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662865, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.354336] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.357624] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.357624] env[65788]: DEBUG nova.compute.manager [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1000.358884] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4c2f9b-fab1-4ac6-a370-b47c7430f59c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.368411] env[65788]: INFO nova.scheduler.client.report [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocations for instance 29c0ab17-5df2-4f6c-bdea-3a4a6c857066 [ 1000.376375] env[65788]: DEBUG nova.compute.manager [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1000.376876] env[65788]: DEBUG nova.objects.instance [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.424150] env[65788]: DEBUG nova.compute.manager [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Received event network-vif-unplugged-9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1000.424150] env[65788]: DEBUG oslo_concurrency.lockutils [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Acquiring lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.424150] env[65788]: DEBUG oslo_concurrency.lockutils [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.424349] env[65788]: DEBUG oslo_concurrency.lockutils [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.424562] env[65788]: DEBUG nova.compute.manager [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] No waiting events found dispatching network-vif-unplugged-9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1000.424709] env[65788]: WARNING nova.compute.manager [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Received unexpected event network-vif-unplugged-9fb26674-54b7-46b3-8502-a80cfdfd36aa for instance with vm_state shelved and task_state shelving_offloading. [ 1000.424911] env[65788]: DEBUG nova.compute.manager [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Received event network-changed-9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1000.425358] env[65788]: DEBUG nova.compute.manager [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Refreshing instance network info cache due to event network-changed-9fb26674-54b7-46b3-8502-a80cfdfd36aa. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1000.425358] env[65788]: DEBUG oslo_concurrency.lockutils [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Acquiring lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.425467] env[65788]: DEBUG oslo_concurrency.lockutils [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Acquired lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1000.425615] env[65788]: DEBUG nova.network.neutron [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Refreshing network info cache for port 9fb26674-54b7-46b3-8502-a80cfdfd36aa {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1000.500705] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081552} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.501214] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.502559] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ce7156-c710-409a-b7ca-c36cf0d1b28d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.535536] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 2bcfbe11-51a2-49b4-b482-02ca332d8c38/2bcfbe11-51a2-49b4-b482-02ca332d8c38.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.536130] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da601114-6728-4818-a013-2e23739b8368 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.565564] env[65788]: DEBUG oslo_vmware.api [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4662871, 'name': PowerOnVM_Task, 'duration_secs': 0.530394} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.567088] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1000.567342] env[65788]: INFO nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1000.567553] env[65788]: DEBUG nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1000.567976] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 1000.567976] env[65788]: value = "task-4662874" [ 1000.567976] env[65788]: _type = "Task" [ 1000.567976] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.568843] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21c8700-a0a3-4449-8042-f57fec24d9eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.586509] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662874, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.592742] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Releasing lock "refresh_cache-4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1000.592742] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Received event network-vif-plugged-20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1000.592742] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquiring lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.592742] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.592742] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.592978] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] No waiting events found dispatching network-vif-plugged-20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1000.593955] env[65788]: WARNING nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Received unexpected event network-vif-plugged-20da233d-2947-43a6-88b1-229b364fbd48 for instance with vm_state building and task_state spawning. [ 1000.593955] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Received event network-changed-20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1000.593955] env[65788]: DEBUG nova.compute.manager [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Refreshing instance network info cache due to event network-changed-20da233d-2947-43a6-88b1-229b364fbd48. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1000.594154] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquiring lock "refresh_cache-2bcfbe11-51a2-49b4-b482-02ca332d8c38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.594262] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Acquired lock "refresh_cache-2bcfbe11-51a2-49b4-b482-02ca332d8c38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1000.594496] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Refreshing network info cache for port 20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1000.650477] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662873, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.650801] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662872, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.669991] env[65788]: DEBUG nova.scheduler.client.report [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 117 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1000.669991] env[65788]: DEBUG nova.compute.provider_tree [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 117 to 118 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1000.670200] env[65788]: DEBUG nova.compute.provider_tree [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.721104] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.721527] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.722400] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.722400] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.722400] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.725951] env[65788]: INFO nova.compute.manager [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Terminating instance [ 1000.819184] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662865, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.871885] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.928798] env[65788]: WARNING neutronclient.v2_0.client [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1000.929548] env[65788]: WARNING openstack [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1000.930176] env[65788]: WARNING openstack [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1000.940552] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.940798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.941094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.941297] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.941469] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.943624] env[65788]: INFO nova.compute.manager [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Terminating instance [ 1001.072610] env[65788]: WARNING openstack [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.072689] env[65788]: WARNING openstack [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.097785] env[65788]: WARNING neutronclient.v2_0.client [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.098423] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.099788] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.107512] env[65788]: INFO nova.compute.manager [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Took 23.03 seconds to build instance. [ 1001.113054] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662874, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.155780] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662872, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.159210] env[65788]: DEBUG oslo_vmware.api [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662873, 'name': PowerOnVM_Task, 'duration_secs': 0.61042} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.159375] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.159683] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537afeb5-8bf7-4472-bfd5-da75412b0de8 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance '7f637326-9388-4d73-a1c7-3e4611ac46c5' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.176555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.869s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.177463] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1001.181460] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.779s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1001.181460] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.184334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.408s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1001.184334] env[65788]: DEBUG nova.objects.instance [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'resources' on Instance uuid a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.230640] env[65788]: INFO nova.scheduler.client.report [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocations for instance 2cb53fe5-aa58-479e-9090-0d8509cf164e [ 1001.232411] env[65788]: DEBUG nova.compute.manager [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1001.233439] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.241182] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61933e71-83bc-45ba-bd8a-dbac0e23e172 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.244867] env[65788]: WARNING neutronclient.v2_0.client [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.245607] env[65788]: WARNING openstack [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.246098] env[65788]: WARNING openstack [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.264264] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.265092] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-81c2262d-12d2-47cc-a656-692d865bb194 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.278390] env[65788]: DEBUG oslo_vmware.api [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 1001.278390] env[65788]: value = "task-4662875" [ 1001.278390] env[65788]: _type = "Task" [ 1001.278390] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.295193] env[65788]: DEBUG oslo_vmware.api [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662875, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.322486] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662865, 'name': CloneVM_Task, 'duration_secs': 1.896871} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.328737] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Created linked-clone VM from snapshot [ 1001.330055] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b504451a-90b5-4b32-888e-4d95c022c043 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.339639] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.340123] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.355639] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Uploading image 17403798-a0dc-480e-ac18-3ae8959fda51 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1001.388554] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.392136] env[65788]: DEBUG nova.network.neutron [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updated VIF entry in instance network info cache for port 9fb26674-54b7-46b3-8502-a80cfdfd36aa. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1001.392136] env[65788]: DEBUG nova.network.neutron [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updating instance_info_cache with network_info: [{"id": "9fb26674-54b7-46b3-8502-a80cfdfd36aa", "address": "fa:16:3e:bd:34:66", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": null, "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9fb26674-54", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1001.393223] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ec9a092-aa34-4b9c-89ed-3658b7505bdb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.398671] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1001.398671] env[65788]: value = "vm-910373" [ 1001.398671] env[65788]: _type = "VirtualMachine" [ 1001.398671] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1001.398964] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e12817a2-3f5d-4eaa-ac9d-3e7c0597a722 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.409761] env[65788]: DEBUG oslo_vmware.api [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1001.409761] env[65788]: value = "task-4662876" [ 1001.409761] env[65788]: _type = "Task" [ 1001.409761] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.412272] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease: (returnval){ [ 1001.412272] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c7d03f-596b-13f2-96e2-38e6d8836237" [ 1001.412272] env[65788]: _type = "HttpNfcLease" [ 1001.412272] env[65788]: } obtained for exporting VM: (result){ [ 1001.412272] env[65788]: value = "vm-910373" [ 1001.412272] env[65788]: _type = "VirtualMachine" [ 1001.412272] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1001.415031] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the lease: (returnval){ [ 1001.415031] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c7d03f-596b-13f2-96e2-38e6d8836237" [ 1001.415031] env[65788]: _type = "HttpNfcLease" [ 1001.415031] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1001.437612] env[65788]: DEBUG oslo_vmware.api [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662876, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.438125] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1001.438125] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c7d03f-596b-13f2-96e2-38e6d8836237" [ 1001.438125] env[65788]: _type = "HttpNfcLease" [ 1001.438125] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1001.447869] env[65788]: DEBUG nova.compute.manager [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1001.448092] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.449593] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0d4134-9978-4730-8110-ab09b8989451 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.463750] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.464168] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc9e67fe-f66d-4782-bc6e-82e0beabc757 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.471823] env[65788]: DEBUG oslo_vmware.api [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 1001.471823] env[65788]: value = "task-4662878" [ 1001.471823] env[65788]: _type = "Task" [ 1001.471823] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.481254] env[65788]: DEBUG oslo_vmware.api [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.489544] env[65788]: WARNING neutronclient.v2_0.client [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.492717] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.492717] env[65788]: WARNING openstack [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.595976] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662874, 'name': ReconfigVM_Task, 'duration_secs': 0.869304} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.596138] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 2bcfbe11-51a2-49b4-b482-02ca332d8c38/2bcfbe11-51a2-49b4-b482-02ca332d8c38.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.597713] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa3e7506-398b-466a-8693-38e01c776641 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.609191] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 1001.609191] env[65788]: value = "task-4662879" [ 1001.609191] env[65788]: _type = "Task" [ 1001.609191] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.615303] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0edb005d-580a-4c37-8589-749a91c0a934 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "1a701758-a056-4948-9069-2a7168f2cc8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.548s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.620615] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662879, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.632273] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Updated VIF entry in instance network info cache for port 20da233d-2947-43a6-88b1-229b364fbd48. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1001.632727] env[65788]: DEBUG nova.network.neutron [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Updating instance_info_cache with network_info: [{"id": "20da233d-2947-43a6-88b1-229b364fbd48", "address": "fa:16:3e:52:cf:5b", "network": {"id": "4e8d7233-937a-4724-93fb-2eb5ff40e13f", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2013035124-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "246c0426008f478aa245d006c3c129eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a58387dd-f438-4913-af6a-fafb734cd881", "external-id": "nsx-vlan-transportzone-169", "segmentation_id": 169, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20da233d-29", "ovs_interfaceid": "20da233d-2947-43a6-88b1-229b364fbd48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1001.647431] env[65788]: DEBUG oslo_vmware.api [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662872, 'name': PowerOnVM_Task, 'duration_secs': 1.467048} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.648487] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1001.648723] env[65788]: INFO nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Took 12.99 seconds to spawn the instance on the hypervisor. [ 1001.648949] env[65788]: DEBUG nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1001.649904] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55777cfe-70b9-4825-99ba-7a12b46257e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.683595] env[65788]: DEBUG nova.compute.utils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1001.686248] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1001.688020] env[65788]: DEBUG nova.network.neutron [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1001.688020] env[65788]: WARNING neutronclient.v2_0.client [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.688020] env[65788]: WARNING neutronclient.v2_0.client [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1001.688361] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1001.688765] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1001.703573] env[65788]: DEBUG nova.objects.instance [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'numa_topology' on Instance uuid a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.759549] env[65788]: DEBUG oslo_concurrency.lockutils [None req-387c05f2-aac6-44da-bd8a-af38c30599f9 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "2cb53fe5-aa58-479e-9090-0d8509cf164e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.866s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1001.788596] env[65788]: DEBUG nova.policy [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b0c1ce04a704a3eaf0a0d4dd09f7e08', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9674e2a5c86b48db8c865a50331ab846', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1001.798177] env[65788]: DEBUG oslo_vmware.api [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662875, 'name': PowerOffVM_Task, 'duration_secs': 0.201464} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.798177] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.798177] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.798511] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3891d1a-ca5a-40b7-8d94-444da42bfd94 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.851358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.875414] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.875775] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.875823] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Deleting the datastore file [datastore2] c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.876187] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c188115-bf13-4f6c-8595-930a9acab77f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.883300] env[65788]: DEBUG oslo_vmware.api [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for the task: (returnval){ [ 1001.883300] env[65788]: value = "task-4662881" [ 1001.883300] env[65788]: _type = "Task" [ 1001.883300] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.892547] env[65788]: DEBUG oslo_vmware.api [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.897259] env[65788]: DEBUG oslo_concurrency.lockutils [req-ebedbe5a-8d95-4d95-b97d-c7d24f91486d req-e3436c87-1635-4236-a40c-e0124f5f3a7e service nova] Releasing lock "refresh_cache-29c0ab17-5df2-4f6c-bdea-3a4a6c857066" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1001.920932] env[65788]: DEBUG oslo_vmware.api [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662876, 'name': PowerOffVM_Task, 'duration_secs': 0.372616} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.924429] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.924638] env[65788]: DEBUG nova.compute.manager [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1001.925414] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91b68d2-1c15-4a60-8bfe-01d47ff5fb08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.931733] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1001.931733] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c7d03f-596b-13f2-96e2-38e6d8836237" [ 1001.931733] env[65788]: _type = "HttpNfcLease" [ 1001.931733] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1001.933522] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1001.933522] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c7d03f-596b-13f2-96e2-38e6d8836237" [ 1001.933522] env[65788]: _type = "HttpNfcLease" [ 1001.933522] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1001.936677] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577401d4-8ca1-4d3d-a533-b2b1407cffbf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.947017] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8d6dd-7551-8260-ab9b-8af6d64407c2/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1001.947126] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8d6dd-7551-8260-ab9b-8af6d64407c2/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1002.021977] env[65788]: DEBUG oslo_vmware.api [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662878, 'name': PowerOffVM_Task, 'duration_secs': 0.328335} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.022370] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.022601] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.022958] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33b0b50e-ed8a-4a80-8e7e-a8ddf325cb26 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.075529] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a3d15758-d0fa-4125-8784-3a1406c44444 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.109471] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.109691] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.109871] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleting the datastore file [datastore1] d78d36e6-a542-4ba8-9e29-b8a7b28c559e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.111191] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c097404f-f0a4-4f91-86f2-ad806c417862 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.133457] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662879, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.135697] env[65788]: DEBUG oslo_concurrency.lockutils [req-791d5893-2828-40f9-8b5d-4816c946e32e req-4c33cdb5-4a76-4e9a-b64c-87afa557c8b4 service nova] Releasing lock "refresh_cache-2bcfbe11-51a2-49b4-b482-02ca332d8c38" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1002.135941] env[65788]: DEBUG oslo_vmware.api [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 1002.135941] env[65788]: value = "task-4662883" [ 1002.135941] env[65788]: _type = "Task" [ 1002.135941] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.145460] env[65788]: DEBUG oslo_vmware.api [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662883, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.182307] env[65788]: INFO nova.compute.manager [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Took 24.72 seconds to build instance. [ 1002.186822] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1002.194482] env[65788]: DEBUG nova.network.neutron [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Successfully created port: 79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1002.206842] env[65788]: DEBUG nova.objects.base [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1002.397864] env[65788]: DEBUG oslo_vmware.api [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Task: {'id': task-4662881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354913} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.398357] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.398357] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.398518] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.399147] env[65788]: INFO nova.compute.manager [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1002.399460] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1002.399683] env[65788]: DEBUG nova.compute.manager [-] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1002.399782] env[65788]: DEBUG nova.network.neutron [-] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1002.400049] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.400613] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.400869] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.443829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f6d17a7d-8e01-4abb-b10d-7ff24c5d3877 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.089s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.580539] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e454e6-6ded-437f-b61e-3eb664ff2d47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.592453] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8301c7-55d6-4495-a20e-c1988d4fe8c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.656201] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6c2dab-d3df-486e-83ca-60cdfb714217 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.670442] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662879, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.680390] env[65788]: DEBUG oslo_vmware.api [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662883, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.399005} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.682427] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.682427] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.682427] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.682427] env[65788]: INFO nova.compute.manager [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1002.682427] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1002.684520] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d958901-9d55-4b9d-80f0-44059c5be57f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.695364] env[65788]: DEBUG nova.compute.manager [-] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1002.696063] env[65788]: DEBUG nova.network.neutron [-] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1002.696531] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1002.697471] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1002.698095] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1002.712710] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a8c8ea9-4dfb-47f9-a0c2-92382db676cb tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.268s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.737718] env[65788]: DEBUG nova.compute.provider_tree [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.152958] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662879, 'name': Rename_Task, 'duration_secs': 1.238167} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.153498] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.153608] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0806ffb6-d334-4cf4-9b29-e3f987a15ff4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.162641] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 1003.162641] env[65788]: value = "task-4662884" [ 1003.162641] env[65788]: _type = "Task" [ 1003.162641] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.173042] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662884, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.221821] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1003.242618] env[65788]: DEBUG nova.scheduler.client.report [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1003.262422] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1003.265780] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1003.266527] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1003.266647] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1003.266887] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1003.267385] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1003.267700] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.267878] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1003.268376] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1003.268376] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1003.268650] env[65788]: DEBUG nova.virt.hardware [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1003.271157] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73fdb46-d25b-4ed1-89d4-0be89254611f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.283143] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441b4e1b-5d55-4ce8-b682-6ac2c97bff65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.584611] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.681048] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662884, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.744878] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1003.748785] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.565s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.751533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.157s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.753025] env[65788]: DEBUG nova.objects.instance [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lazy-loading 'resources' on Instance uuid af9120a6-2cea-46da-ba06-6036b9beda78 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.909032] env[65788]: DEBUG nova.network.neutron [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Successfully updated port: 79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1004.183155] env[65788]: DEBUG oslo_vmware.api [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662884, 'name': PowerOnVM_Task, 'duration_secs': 0.693008} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.183483] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.186111] env[65788]: INFO nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Took 8.84 seconds to spawn the instance on the hypervisor. [ 1004.186569] env[65788]: DEBUG nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1004.187368] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d68b57a-3b45-4652-a512-5b053fba9009 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.268239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-56b56515-2412-4a75-8d88-2d69c3e64b86 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 39.538s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.270784] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 16.355s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.271088] env[65788]: INFO nova.compute.manager [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Unshelving [ 1004.414891] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "refresh_cache-902d8a69-0398-4752-a609-0ac5b4b32e27" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.415256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "refresh_cache-902d8a69-0398-4752-a609-0ac5b4b32e27" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1004.415503] env[65788]: DEBUG nova.network.neutron [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1004.610585] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd84d5e-983c-4431-886c-a44b35f2cb44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.623818] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79bbf564-1b53-402b-8bc0-6e185ffd6ef5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.629499] env[65788]: DEBUG nova.network.neutron [-] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.666040] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbc58e8-fcfa-4d7f-ae0a-9a416d434790 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.676341] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7f735a-0b0b-4f14-96ad-f1839ebdade6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.693651] env[65788]: DEBUG nova.compute.provider_tree [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.714041] env[65788]: INFO nova.compute.manager [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Took 26.20 seconds to build instance. [ 1004.765581] env[65788]: DEBUG nova.compute.manager [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Received event network-changed-9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1004.765804] env[65788]: DEBUG nova.compute.manager [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Refreshing instance network info cache due to event network-changed-9f372d9b-f11d-40fc-b795-bae66f4a1800. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1004.766096] env[65788]: DEBUG oslo_concurrency.lockutils [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Acquiring lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.766096] env[65788]: DEBUG oslo_concurrency.lockutils [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Acquired lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1004.766259] env[65788]: DEBUG nova.network.neutron [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Refreshing network info cache for port 9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1004.910035] env[65788]: DEBUG nova.network.neutron [-] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1004.920033] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.920155] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.959562] env[65788]: DEBUG nova.network.neutron [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1004.979664] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1004.980095] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1004.991579] env[65788]: DEBUG nova.compute.manager [req-e5083822-ee2c-4ca2-b19d-16ff542a0846 req-0dc43d9b-cb31-4b5b-9a8e-023b5700aab3 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Received event network-vif-plugged-79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1004.991974] env[65788]: DEBUG oslo_concurrency.lockutils [req-e5083822-ee2c-4ca2-b19d-16ff542a0846 req-0dc43d9b-cb31-4b5b-9a8e-023b5700aab3 service nova] Acquiring lock "902d8a69-0398-4752-a609-0ac5b4b32e27-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.992074] env[65788]: DEBUG oslo_concurrency.lockutils [req-e5083822-ee2c-4ca2-b19d-16ff542a0846 req-0dc43d9b-cb31-4b5b-9a8e-023b5700aab3 service nova] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.992178] env[65788]: DEBUG oslo_concurrency.lockutils [req-e5083822-ee2c-4ca2-b19d-16ff542a0846 req-0dc43d9b-cb31-4b5b-9a8e-023b5700aab3 service nova] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.992345] env[65788]: DEBUG nova.compute.manager [req-e5083822-ee2c-4ca2-b19d-16ff542a0846 req-0dc43d9b-cb31-4b5b-9a8e-023b5700aab3 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] No waiting events found dispatching network-vif-plugged-79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1004.992505] env[65788]: WARNING nova.compute.manager [req-e5083822-ee2c-4ca2-b19d-16ff542a0846 req-0dc43d9b-cb31-4b5b-9a8e-023b5700aab3 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Received unexpected event network-vif-plugged-79eedfab-2cdd-4443-821e-f2a69d774f98 for instance with vm_state building and task_state spawning. [ 1005.066960] env[65788]: WARNING neutronclient.v2_0.client [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.067673] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.068132] env[65788]: WARNING openstack [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.135961] env[65788]: INFO nova.compute.manager [-] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Took 2.74 seconds to deallocate network for instance. [ 1005.197378] env[65788]: DEBUG nova.scheduler.client.report [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1005.207708] env[65788]: DEBUG nova.network.neutron [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Updating instance_info_cache with network_info: [{"id": "79eedfab-2cdd-4443-821e-f2a69d774f98", "address": "fa:16:3e:c3:a9:32", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79eedfab-2c", "ovs_interfaceid": "79eedfab-2cdd-4443-821e-f2a69d774f98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1005.217507] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10a9758d-3b28-4353-81cc-9c960a1fa219 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.720s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.269024] env[65788]: WARNING neutronclient.v2_0.client [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.272183] env[65788]: WARNING openstack [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.272183] env[65788]: WARNING openstack [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.302751] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.411922] env[65788]: INFO nova.compute.manager [-] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Took 2.72 seconds to deallocate network for instance. [ 1005.458646] env[65788]: WARNING openstack [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.459113] env[65788]: WARNING openstack [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.609534] env[65788]: WARNING neutronclient.v2_0.client [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.610136] env[65788]: WARNING openstack [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1005.610574] env[65788]: WARNING openstack [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1005.640121] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.649354] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.703773] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.952s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.707179] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.965s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.710065] env[65788]: INFO nova.compute.claims [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.713830] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.714807] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1005.717373] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "refresh_cache-902d8a69-0398-4752-a609-0ac5b4b32e27" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1005.719989] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Instance network_info: |[{"id": "79eedfab-2cdd-4443-821e-f2a69d774f98", "address": "fa:16:3e:c3:a9:32", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79eedfab-2c", "ovs_interfaceid": "79eedfab-2cdd-4443-821e-f2a69d774f98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1005.719989] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:a9:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '980cb890-345b-4bf8-990a-a2faec78e49c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79eedfab-2cdd-4443-821e-f2a69d774f98', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1005.728043] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1005.728285] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1005.728600] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2953602a-6c82-4e65-b187-46e3d51409d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.749492] env[65788]: INFO nova.scheduler.client.report [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Deleted allocations for instance af9120a6-2cea-46da-ba06-6036b9beda78 [ 1005.757557] env[65788]: DEBUG nova.objects.instance [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.758860] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1005.758860] env[65788]: value = "task-4662885" [ 1005.758860] env[65788]: _type = "Task" [ 1005.758860] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.769562] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662885, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.835316] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.835845] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.835845] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.836042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.839018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.839018] env[65788]: INFO nova.compute.manager [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Terminating instance [ 1005.868288] env[65788]: DEBUG nova.network.neutron [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Updated VIF entry in instance network info cache for port 9f372d9b-f11d-40fc-b795-bae66f4a1800. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1005.869193] env[65788]: DEBUG nova.network.neutron [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Updating instance_info_cache with network_info: [{"id": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "address": "fa:16:3e:24:47:96", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f372d9b-f1", "ovs_interfaceid": "9f372d9b-f11d-40fc-b795-bae66f4a1800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1005.902792] env[65788]: DEBUG nova.network.neutron [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Port e98f1b15-0a05-4be4-b950-dd7e6749eb8f binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1005.903181] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.903726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1005.903726] env[65788]: DEBUG nova.network.neutron [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1005.918269] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.265725] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ee50d427-ab46-437d-a0f0-ea4239868bdc tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "af9120a6-2cea-46da-ba06-6036b9beda78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.142s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1006.271091] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.271497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1006.271595] env[65788]: DEBUG nova.network.neutron [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1006.272249] env[65788]: DEBUG nova.objects.instance [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'info_cache' on Instance uuid d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.280880] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662885, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.342968] env[65788]: DEBUG nova.compute.manager [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1006.343359] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.344448] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a44744-329b-4b72-bceb-7eada63da01e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.355314] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.355685] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-84521046-62c1-49d3-8127-a2f0f124e0ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.364687] env[65788]: DEBUG oslo_vmware.api [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 1006.364687] env[65788]: value = "task-4662886" [ 1006.364687] env[65788]: _type = "Task" [ 1006.364687] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.372447] env[65788]: DEBUG oslo_concurrency.lockutils [req-76f066f8-ad81-4b84-8f83-0e0cb84cdd0b req-6fd9a820-a7dc-4ad5-96db-da2de3fdcc7a service nova] Releasing lock "refresh_cache-1a701758-a056-4948-9069-2a7168f2cc8c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1006.380415] env[65788]: DEBUG oslo_vmware.api [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662886, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.407180] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.407983] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.408420] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.605201] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.605201] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.683826] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.684763] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1006.685255] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1006.694848] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.695628] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.695628] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.695894] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.696159] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1006.698502] env[65788]: INFO nova.compute.manager [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Terminating instance [ 1006.776417] env[65788]: DEBUG nova.objects.base [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1006.782483] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662885, 'name': CreateVM_Task, 'duration_secs': 0.559025} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.782922] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1006.783468] env[65788]: WARNING neutronclient.v2_0.client [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1006.783831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.783982] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1006.784323] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1006.784620] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1e4deef-04b5-4083-828d-d0292cec7894 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.792507] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1006.792507] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b00076-c3d9-d938-a180-523d17278a86" [ 1006.792507] env[65788]: _type = "Task" [ 1006.792507] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.807681] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b00076-c3d9-d938-a180-523d17278a86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.810961] env[65788]: DEBUG nova.network.neutron [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1006.884785] env[65788]: DEBUG oslo_vmware.api [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662886, 'name': PowerOffVM_Task, 'duration_secs': 0.33195} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.889025] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.889965] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1006.891921] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90091a0a-f95e-4693-be0d-206f01e927a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.026380] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.026380] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.026577] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Deleting the datastore file [datastore2] 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.030038] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a672da51-5b69-4a45-9e13-e9de8345833f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.039518] env[65788]: DEBUG oslo_vmware.api [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for the task: (returnval){ [ 1007.039518] env[65788]: value = "task-4662889" [ 1007.039518] env[65788]: _type = "Task" [ 1007.039518] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.050373] env[65788]: DEBUG oslo_vmware.api [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.070584] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f3210f-e195-4a89-b715-2f8a99b11aee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.081778] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c9ada3-dae4-4873-ad45-6a60f9d11058 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.121302] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0732458e-c211-4e2d-a0d6-f90b00ac70d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.131301] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7837dba9-6f96-4adb-88d8-66d0c5e7e53b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.147354] env[65788]: DEBUG nova.compute.provider_tree [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.206038] env[65788]: DEBUG nova.compute.manager [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1007.206676] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.207556] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff16cbe-7b4e-46c1-b06c-36016020dbfb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.217612] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.217992] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a395bf5-67c8-4855-a5b1-b6bb6b669600 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.226815] env[65788]: DEBUG oslo_vmware.api [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 1007.226815] env[65788]: value = "task-4662890" [ 1007.226815] env[65788]: _type = "Task" [ 1007.226815] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.240656] env[65788]: DEBUG oslo_vmware.api [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.263660] env[65788]: DEBUG nova.compute.manager [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Received event network-changed-79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1007.264900] env[65788]: DEBUG nova.compute.manager [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Refreshing instance network info cache due to event network-changed-79eedfab-2cdd-4443-821e-f2a69d774f98. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1007.264900] env[65788]: DEBUG oslo_concurrency.lockutils [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Acquiring lock "refresh_cache-902d8a69-0398-4752-a609-0ac5b4b32e27" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.264900] env[65788]: DEBUG oslo_concurrency.lockutils [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Acquired lock "refresh_cache-902d8a69-0398-4752-a609-0ac5b4b32e27" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.264900] env[65788]: DEBUG nova.network.neutron [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Refreshing network info cache for port 79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1007.284047] env[65788]: WARNING neutronclient.v2_0.client [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.284974] env[65788]: WARNING openstack [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.285557] env[65788]: WARNING openstack [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.307129] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b00076-c3d9-d938-a180-523d17278a86, 'name': SearchDatastore_Task, 'duration_secs': 0.020346} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.307588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1007.307874] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1007.308207] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.308386] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.308615] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1007.309618] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd89734c-cdbd-495d-8a7e-234d0aaa8fa4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.316662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1007.320663] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1007.320917] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1007.322148] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be46236f-245b-4155-935a-ae73820ea867 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.330063] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1007.330063] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521e8ec4-24e5-1792-bf36-3f5e2ac430a3" [ 1007.330063] env[65788]: _type = "Task" [ 1007.330063] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.341392] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521e8ec4-24e5-1792-bf36-3f5e2ac430a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.448851] env[65788]: WARNING openstack [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.449184] env[65788]: WARNING openstack [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.532328] env[65788]: WARNING neutronclient.v2_0.client [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.533092] env[65788]: WARNING openstack [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.533464] env[65788]: WARNING openstack [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.555934] env[65788]: DEBUG oslo_vmware.api [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Task: {'id': task-4662889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326908} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.556316] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1007.556549] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1007.556803] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1007.557314] env[65788]: INFO nova.compute.manager [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1007.558015] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1007.558386] env[65788]: DEBUG nova.compute.manager [-] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1007.558650] env[65788]: DEBUG nova.network.neutron [-] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1007.559055] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.559747] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.560164] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.603086] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.603422] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.651842] env[65788]: DEBUG nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1007.668401] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.720625] env[65788]: DEBUG nova.network.neutron [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [{"id": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "address": "fa:16:3e:88:8c:66", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ac3e7fa-b2", "ovs_interfaceid": "4ac3e7fa-b26e-4239-be70-ae1397fbba52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1007.737940] env[65788]: DEBUG oslo_vmware.api [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662890, 'name': PowerOffVM_Task, 'duration_secs': 0.217318} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.738227] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.738408] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.738661] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3261fba3-dcc2-4116-98d3-34f9a96a5fac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.767519] env[65788]: WARNING neutronclient.v2_0.client [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1007.768436] env[65788]: WARNING openstack [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.768780] env[65788]: WARNING openstack [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1007.808194] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.811193] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.811193] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleting the datastore file [datastore2] 2bcfbe11-51a2-49b4-b482-02ca332d8c38 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.811193] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59beed18-0830-43bd-a39a-930db1db4f80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.817294] env[65788]: DEBUG oslo_vmware.api [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for the task: (returnval){ [ 1007.817294] env[65788]: value = "task-4662892" [ 1007.817294] env[65788]: _type = "Task" [ 1007.817294] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.822401] env[65788]: DEBUG nova.compute.manager [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65788) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:925}} [ 1007.830471] env[65788]: DEBUG oslo_vmware.api [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662892, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.843440] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521e8ec4-24e5-1792-bf36-3f5e2ac430a3, 'name': SearchDatastore_Task, 'duration_secs': 0.016719} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.844361] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a87c54a-fc08-45b6-97cb-463fa4cbce1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.856032] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1007.856032] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524eeb2c-6dee-d23b-c464-c90a9a9548c0" [ 1007.856032] env[65788]: _type = "Task" [ 1007.856032] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.865910] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524eeb2c-6dee-d23b-c464-c90a9a9548c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.974452] env[65788]: WARNING openstack [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1007.975098] env[65788]: WARNING openstack [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.081883] env[65788]: WARNING neutronclient.v2_0.client [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.082730] env[65788]: WARNING openstack [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.083216] env[65788]: WARNING openstack [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.106662] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1008.156859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.157738] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1008.160524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.345s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.162957] env[65788]: INFO nova.compute.claims [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.223298] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "refresh_cache-d8103991-fd1f-4a57-81ce-1a47dc4defe0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.262444] env[65788]: DEBUG nova.network.neutron [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Updated VIF entry in instance network info cache for port 79eedfab-2cdd-4443-821e-f2a69d774f98. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1008.263119] env[65788]: DEBUG nova.network.neutron [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Updating instance_info_cache with network_info: [{"id": "79eedfab-2cdd-4443-821e-f2a69d774f98", "address": "fa:16:3e:c3:a9:32", "network": {"id": "5f742516-7eb8-42c9-838d-11394d520333", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1782550395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "9674e2a5c86b48db8c865a50331ab846", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "980cb890-345b-4bf8-990a-a2faec78e49c", "external-id": "nsx-vlan-transportzone-965", "segmentation_id": 965, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79eedfab-2c", "ovs_interfaceid": "79eedfab-2cdd-4443-821e-f2a69d774f98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.332843] env[65788]: DEBUG oslo_vmware.api [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Task: {'id': task-4662892, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.396036} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.332843] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.332843] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.332843] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.332843] env[65788]: INFO nova.compute.manager [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1008.332843] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1008.333570] env[65788]: DEBUG nova.compute.manager [-] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1008.333570] env[65788]: DEBUG nova.network.neutron [-] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1008.333570] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.334159] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.334403] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.368574] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524eeb2c-6dee-d23b-c464-c90a9a9548c0, 'name': SearchDatastore_Task, 'duration_secs': 0.023403} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.369618] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.370147] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 902d8a69-0398-4752-a609-0ac5b4b32e27/902d8a69-0398-4752-a609-0ac5b4b32e27.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1008.374109] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5d9f83b-0e7a-4e04-94ce-f8210df8e4ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.381891] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1008.381891] env[65788]: value = "task-4662893" [ 1008.381891] env[65788]: _type = "Task" [ 1008.381891] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.398736] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.423812] env[65788]: DEBUG nova.compute.manager [req-fba1c855-2a19-454f-b224-afc22ff30cf1 req-ef67c19a-14db-41ab-aa28-d2f5ee77286a service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received event network-vif-deleted-7b27f93c-6a27-456b-b3aa-04d0cc66c075 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1008.423812] env[65788]: INFO nova.compute.manager [req-fba1c855-2a19-454f-b224-afc22ff30cf1 req-ef67c19a-14db-41ab-aa28-d2f5ee77286a service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Neutron deleted interface 7b27f93c-6a27-456b-b3aa-04d0cc66c075; detaching it from the instance and deleting it from the info cache [ 1008.423812] env[65788]: DEBUG nova.network.neutron [req-fba1c855-2a19-454f-b224-afc22ff30cf1 req-ef67c19a-14db-41ab-aa28-d2f5ee77286a service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updating instance_info_cache with network_info: [{"id": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "address": "fa:16:3e:e0:3f:54", "network": {"id": "3845b505-1663-4cb0-88cc-f080dee79756", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2071292552", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd2c665795dc4241a50c46de045f2c80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5413129-8b", "ovs_interfaceid": "e5413129-8bf0-444f-9791-900f8e5c2f6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.427447] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.638695] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.668754] env[65788]: DEBUG nova.compute.utils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1008.673781] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1008.673781] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1008.674124] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.674559] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1008.675437] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1008.676084] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1008.753271] env[65788]: DEBUG nova.network.neutron [-] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1008.761257] env[65788]: DEBUG nova.policy [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72684a9d23904034af7dcda5026dcf93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcd247e1f9bb4cc5b29e009f727bacca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1008.770420] env[65788]: DEBUG oslo_concurrency.lockutils [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] Releasing lock "refresh_cache-902d8a69-0398-4752-a609-0ac5b4b32e27" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.770420] env[65788]: DEBUG nova.compute.manager [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Received event network-vif-deleted-6e2a5ea0-49ef-4baf-8210-ab0ca08a7d48 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1008.770766] env[65788]: DEBUG nova.compute.manager [req-efa3df1e-1cb3-46a5-94ee-14c417476480 req-58854d9d-c76b-4bd1-bc35-af420c1d2186 service nova] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Received event network-vif-deleted-2af7b9f3-396f-41c3-954c-e947acc02fd1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1008.899515] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662893, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.930330] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07a4a856-f369-425f-82f5-c34f89e217bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.950540] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a8a5aa-d995-49a5-92be-09f12093df6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.972193] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.996147] env[65788]: DEBUG nova.compute.manager [req-fba1c855-2a19-454f-b224-afc22ff30cf1 req-ef67c19a-14db-41ab-aa28-d2f5ee77286a service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Detach interface failed, port_id=7b27f93c-6a27-456b-b3aa-04d0cc66c075, reason: Instance 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1009.097649] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfccfea-e8b7-4cd4-876b-0dd8394879d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.110467] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834a73a7-a0c7-4521-bca0-6c8a4c1a8052 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.154582] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50d24a7-9ba3-4587-b1c1-232922ce6d1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.158576] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Successfully created port: bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1009.168519] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a50490-cb8a-4103-8120-d8515d573b88 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.190821] env[65788]: DEBUG nova.compute.provider_tree [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.192625] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1009.212374] env[65788]: DEBUG nova.network.neutron [-] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1009.230712] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1009.231108] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d27aa0ce-18f5-46b3-8a4f-b0c2fbd33809 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.241396] env[65788]: DEBUG oslo_vmware.api [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1009.241396] env[65788]: value = "task-4662894" [ 1009.241396] env[65788]: _type = "Task" [ 1009.241396] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.251928] env[65788]: DEBUG oslo_vmware.api [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662894, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.257086] env[65788]: INFO nova.compute.manager [-] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Took 1.70 seconds to deallocate network for instance. [ 1009.390482] env[65788]: DEBUG nova.compute.manager [req-cf5d3287-1497-4f54-ac52-2277fde7a6b1 req-8af258db-76b8-4a1b-858c-ce84f62d38de service nova] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Received event network-vif-deleted-e5413129-8bf0-444f-9791-900f8e5c2f6e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1009.390787] env[65788]: DEBUG nova.compute.manager [req-cf5d3287-1497-4f54-ac52-2277fde7a6b1 req-8af258db-76b8-4a1b-858c-ce84f62d38de service nova] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Received event network-vif-deleted-20da233d-2947-43a6-88b1-229b364fbd48 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1009.408733] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662893, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.725045} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.408733] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 902d8a69-0398-4752-a609-0ac5b4b32e27/902d8a69-0398-4752-a609-0ac5b4b32e27.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1009.408733] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.409044] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76ddcbd5-96e3-4655-a7d0-f72625993437 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.418441] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1009.418441] env[65788]: value = "task-4662895" [ 1009.418441] env[65788]: _type = "Task" [ 1009.418441] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.428014] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662895, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.528263] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "374564df-70fa-4b89-8f38-e559245e5ebf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.528551] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "374564df-70fa-4b89-8f38-e559245e5ebf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.528910] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "374564df-70fa-4b89-8f38-e559245e5ebf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.529048] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "374564df-70fa-4b89-8f38-e559245e5ebf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1009.529280] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "374564df-70fa-4b89-8f38-e559245e5ebf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1009.532037] env[65788]: INFO nova.compute.manager [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Terminating instance [ 1009.715168] env[65788]: INFO nova.compute.manager [-] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Took 1.38 seconds to deallocate network for instance. [ 1009.724639] env[65788]: ERROR nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [req-a69c89c5-1570-4cc4-afa1-ce7a4b750d58] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a69c89c5-1570-4cc4-afa1-ce7a4b750d58"}]} [ 1009.749087] env[65788]: DEBUG nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1009.755307] env[65788]: DEBUG oslo_vmware.api [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662894, 'name': PowerOnVM_Task, 'duration_secs': 0.475469} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.755678] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.755832] env[65788]: DEBUG nova.compute.manager [None req-fb62d973-24d6-4926-a236-f41541ef6ea2 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1009.757069] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b6ea5b-0399-4544-9d55-9c5dbf65d215 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.764099] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1009.770183] env[65788]: DEBUG nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1009.770435] env[65788]: DEBUG nova.compute.provider_tree [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.790778] env[65788]: DEBUG nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1009.820340] env[65788]: DEBUG nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1009.929166] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662895, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077104} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.932770] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1009.934183] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e84007a-fa9d-43cf-98f4-ba74210477ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.959731] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 902d8a69-0398-4752-a609-0ac5b4b32e27/902d8a69-0398-4752-a609-0ac5b4b32e27.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.962693] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-851cbe90-279e-4db9-9732-ddb016fe9a30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.984433] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1009.984433] env[65788]: value = "task-4662896" [ 1009.984433] env[65788]: _type = "Task" [ 1009.984433] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.999481] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662896, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.040029] env[65788]: DEBUG nova.compute.manager [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1010.040029] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.040029] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2580d7-1081-401a-ab34-fa49cbbcf712 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.050725] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.051362] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdbe7638-29a5-4f10-be3e-3fd226a82702 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.060054] env[65788]: DEBUG oslo_vmware.api [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 1010.060054] env[65788]: value = "task-4662897" [ 1010.060054] env[65788]: _type = "Task" [ 1010.060054] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.073197] env[65788]: DEBUG oslo_vmware.api [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662897, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.163487] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d39d8dd-d9bc-4053-8024-ec0918fe34cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.174967] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8423bed6-6bb5-419a-9351-fadeb6515128 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.235040] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1010.238156] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.240061] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3441d6e2-41f5-42fc-9334-7e1384fd94ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.251463] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec526ffc-795f-4e17-a1b8-513fb3ca8712 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.271078] env[65788]: DEBUG nova.compute.provider_tree [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.275142] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1010.275473] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1010.275648] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1010.275838] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1010.276022] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1010.276191] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1010.276438] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.276601] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1010.276954] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1010.276954] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1010.277165] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1010.280496] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f206321a-18fd-4841-8e1b-e157845744e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.296926] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068799b1-8826-4c25-b297-64cdb65d785b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.326987] env[65788]: DEBUG nova.scheduler.client.report [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1010.327391] env[65788]: DEBUG nova.compute.provider_tree [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 119 to 120 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1010.327684] env[65788]: DEBUG nova.compute.provider_tree [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.496076] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.571085] env[65788]: DEBUG oslo_vmware.api [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662897, 'name': PowerOffVM_Task, 'duration_secs': 0.37438} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.571384] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.571550] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.571821] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a5de0ca-d74a-48ac-ba7b-eaf1f0fc1bc0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.658420] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.658775] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.659082] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Deleting the datastore file [datastore2] 374564df-70fa-4b89-8f38-e559245e5ebf {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.659515] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58fb2e86-7b25-4522-bc94-c58c04ffe300 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.670456] env[65788]: DEBUG oslo_vmware.api [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for the task: (returnval){ [ 1010.670456] env[65788]: value = "task-4662900" [ 1010.670456] env[65788]: _type = "Task" [ 1010.670456] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.680682] env[65788]: DEBUG oslo_vmware.api [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662900, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.835867] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.674s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.835867] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1010.837779] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.966s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.838083] env[65788]: DEBUG nova.objects.instance [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'resources' on Instance uuid 29c0ab17-5df2-4f6c-bdea-3a4a6c857066 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.865622] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Successfully updated port: bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1010.997564] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662896, 'name': ReconfigVM_Task, 'duration_secs': 0.570642} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.997929] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 902d8a69-0398-4752-a609-0ac5b4b32e27/902d8a69-0398-4752-a609-0ac5b4b32e27.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.998636] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9a76281-d1d0-4a6a-9e4c-87fe5b619cc9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.006716] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1011.006716] env[65788]: value = "task-4662901" [ 1011.006716] env[65788]: _type = "Task" [ 1011.006716] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.017147] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662901, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.181744] env[65788]: DEBUG oslo_vmware.api [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Task: {'id': task-4662900, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.445043} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.182073] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.182304] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.182549] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.182771] env[65788]: INFO nova.compute.manager [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1011.183124] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1011.183367] env[65788]: DEBUG nova.compute.manager [-] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1011.183472] env[65788]: DEBUG nova.network.neutron [-] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1011.183741] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.184351] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.184627] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.260571] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.341652] env[65788]: DEBUG nova.compute.utils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1011.343677] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1011.343901] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1011.344245] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.344562] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1011.345169] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.345514] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.354678] env[65788]: DEBUG nova.objects.instance [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'numa_topology' on Instance uuid 29c0ab17-5df2-4f6c-bdea-3a4a6c857066 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.367858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "refresh_cache-6e78042d-ba10-4629-816f-3a13b2e22d4e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.368076] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "refresh_cache-6e78042d-ba10-4629-816f-3a13b2e22d4e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1011.368258] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1011.394969] env[65788]: DEBUG nova.policy [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72684a9d23904034af7dcda5026dcf93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcd247e1f9bb4cc5b29e009f727bacca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1011.427990] env[65788]: DEBUG nova.compute.manager [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Received event network-vif-plugged-bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1011.428257] env[65788]: DEBUG oslo_concurrency.lockutils [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Acquiring lock "6e78042d-ba10-4629-816f-3a13b2e22d4e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.428504] env[65788]: DEBUG oslo_concurrency.lockutils [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.428670] env[65788]: DEBUG oslo_concurrency.lockutils [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.428838] env[65788]: DEBUG nova.compute.manager [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] No waiting events found dispatching network-vif-plugged-bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1011.429074] env[65788]: WARNING nova.compute.manager [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Received unexpected event network-vif-plugged-bceffc39-3195-4064-9ae1-646b7a191ae7 for instance with vm_state building and task_state spawning. [ 1011.429250] env[65788]: DEBUG nova.compute.manager [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Received event network-changed-bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1011.429443] env[65788]: DEBUG nova.compute.manager [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Refreshing instance network info cache due to event network-changed-bceffc39-3195-4064-9ae1-646b7a191ae7. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1011.429674] env[65788]: DEBUG oslo_concurrency.lockutils [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Acquiring lock "refresh_cache-6e78042d-ba10-4629-816f-3a13b2e22d4e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.518870] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662901, 'name': Rename_Task, 'duration_secs': 0.274423} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.519348] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.519348] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-559bf1bd-f5bf-4ee0-8c1c-ae2824a8b369 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.529490] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1011.529490] env[65788]: value = "task-4662902" [ 1011.529490] env[65788]: _type = "Task" [ 1011.529490] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.539788] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662902, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.581361] env[65788]: DEBUG nova.compute.manager [req-1fb99db2-b27b-4fa1-9863-ac0d512e1371 req-e4be23df-45b1-4333-bbf7-712704220ca0 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Received event network-vif-deleted-85fbe51b-2508-45d1-a3e3-b6c38fac2060 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1011.581361] env[65788]: INFO nova.compute.manager [req-1fb99db2-b27b-4fa1-9863-ac0d512e1371 req-e4be23df-45b1-4333-bbf7-712704220ca0 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Neutron deleted interface 85fbe51b-2508-45d1-a3e3-b6c38fac2060; detaching it from the instance and deleting it from the info cache [ 1011.581361] env[65788]: DEBUG nova.network.neutron [req-1fb99db2-b27b-4fa1-9863-ac0d512e1371 req-e4be23df-45b1-4333-bbf7-712704220ca0 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1011.712924] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Successfully created port: 81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1011.853597] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1011.857483] env[65788]: DEBUG nova.objects.base [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Object Instance<29c0ab17-5df2-4f6c-bdea-3a4a6c857066> lazy-loaded attributes: resources,numa_topology {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1011.870742] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.871369] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.938166] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1011.983577] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1011.984190] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1011.996340] env[65788]: DEBUG nova.network.neutron [-] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1012.030238] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8d6dd-7551-8260-ab9b-8af6d64407c2/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1012.034711] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee3bb43-778d-41b9-8b40-41704442f7ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.046819] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662902, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.051623] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8d6dd-7551-8260-ab9b-8af6d64407c2/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1012.052317] env[65788]: ERROR oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8d6dd-7551-8260-ab9b-8af6d64407c2/disk-0.vmdk due to incomplete transfer. [ 1012.052317] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-71e1b6b7-6201-4867-a055-bdbf3d605155 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.062173] env[65788]: DEBUG oslo_vmware.rw_handles [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d8d6dd-7551-8260-ab9b-8af6d64407c2/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1012.062347] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Uploaded image 17403798-a0dc-480e-ac18-3ae8959fda51 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1012.065374] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1012.068772] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-27639dda-f696-42d2-9f33-17dbbd3ea504 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.072584] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1012.073292] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1012.073686] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1012.087777] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1aa89756-9bf4-4421-9d19-19caf9fcc169 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.095957] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1012.095957] env[65788]: value = "task-4662903" [ 1012.095957] env[65788]: _type = "Task" [ 1012.095957] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.107372] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a70a3dc-58d0-4edd-ac2d-15bde8163190 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.126170] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662903, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.150500] env[65788]: DEBUG nova.compute.manager [req-1fb99db2-b27b-4fa1-9863-ac0d512e1371 req-e4be23df-45b1-4333-bbf7-712704220ca0 service nova] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Detach interface failed, port_id=85fbe51b-2508-45d1-a3e3-b6c38fac2060, reason: Instance 374564df-70fa-4b89-8f38-e559245e5ebf could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1012.247456] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Updating instance_info_cache with network_info: [{"id": "bceffc39-3195-4064-9ae1-646b7a191ae7", "address": "fa:16:3e:6a:ad:3e", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbceffc39-31", "ovs_interfaceid": "bceffc39-3195-4064-9ae1-646b7a191ae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1012.323963] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d7d6c0-ba28-47d4-8cfe-b7e5b419408b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.334769] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4bf2fb-ccdf-4572-9e58-225fbc1259f6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.374578] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5d6ee3-265d-49f9-bcfa-bf3276edf41a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.383666] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc0b0c8-8052-4d54-a680-d39a24186b11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.400973] env[65788]: DEBUG nova.compute.provider_tree [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.501296] env[65788]: INFO nova.compute.manager [-] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Took 1.32 seconds to deallocate network for instance. [ 1012.542969] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662902, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.610191] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662903, 'name': Destroy_Task, 'duration_secs': 0.398826} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.610191] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Destroyed the VM [ 1012.610274] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1012.610500] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a649e126-5a70-40b3-b5fe-ebbc15a42448 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.619366] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1012.619366] env[65788]: value = "task-4662905" [ 1012.619366] env[65788]: _type = "Task" [ 1012.619366] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.632114] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662905, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.751260] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "refresh_cache-6e78042d-ba10-4629-816f-3a13b2e22d4e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1012.751754] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Instance network_info: |[{"id": "bceffc39-3195-4064-9ae1-646b7a191ae7", "address": "fa:16:3e:6a:ad:3e", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbceffc39-31", "ovs_interfaceid": "bceffc39-3195-4064-9ae1-646b7a191ae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1012.752186] env[65788]: DEBUG oslo_concurrency.lockutils [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Acquired lock "refresh_cache-6e78042d-ba10-4629-816f-3a13b2e22d4e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1012.752450] env[65788]: DEBUG nova.network.neutron [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Refreshing network info cache for port bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1012.754042] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:ad:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bceffc39-3195-4064-9ae1-646b7a191ae7', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.764304] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Creating folder: Project (fcd247e1f9bb4cc5b29e009f727bacca). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.765622] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2edeb165-9c5c-473c-8564-b9b1f2f1e716 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.779396] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Created folder: Project (fcd247e1f9bb4cc5b29e009f727bacca) in parent group-v910111. [ 1012.779652] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Creating folder: Instances. Parent ref: group-v910376. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.779961] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b2b54db-d678-4291-9c4e-04d797503b13 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.792857] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Created folder: Instances in parent group-v910376. [ 1012.793166] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1012.793399] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1012.793663] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0964c98-255b-4fb4-8063-9af4e35dbb04 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.815813] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.815813] env[65788]: value = "task-4662908" [ 1012.815813] env[65788]: _type = "Task" [ 1012.815813] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.825021] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662908, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.879153] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1012.907351] env[65788]: DEBUG nova.scheduler.client.report [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1012.914532] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1012.914759] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1012.914918] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1012.915333] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1012.915333] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1012.915469] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1012.915648] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.915813] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1012.915977] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1012.916158] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1012.916336] env[65788]: DEBUG nova.virt.hardware [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1012.917287] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948ff1bb-e295-403d-8153-9a2d1664f92e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.926534] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7b0c70-fe2e-4857-bc28-3e14c7e03f2e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.009844] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.043816] env[65788]: DEBUG oslo_vmware.api [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662902, 'name': PowerOnVM_Task, 'duration_secs': 1.486072} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.044055] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.044290] env[65788]: INFO nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Took 9.82 seconds to spawn the instance on the hypervisor. [ 1013.044472] env[65788]: DEBUG nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1013.045320] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022dafc7-fe48-4f66-8782-93bd51d6a58e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.132738] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662905, 'name': RemoveSnapshot_Task, 'duration_secs': 0.481079} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.133071] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1013.133442] env[65788]: DEBUG nova.compute.manager [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1013.134275] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6389fb0c-93d3-4b1a-90c2-248488369547 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.266129] env[65788]: WARNING neutronclient.v2_0.client [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.267035] env[65788]: WARNING openstack [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.267744] env[65788]: WARNING openstack [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.283299] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Successfully updated port: 81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1013.328152] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662908, 'name': CreateVM_Task, 'duration_secs': 0.450911} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.328152] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.328152] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.328483] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.328483] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.329217] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1013.329217] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13c77436-7776-4881-8b70-d7bdc5ac87ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.335615] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1013.335615] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218ebae-7a18-8a87-8d3e-7902850d9b70" [ 1013.335615] env[65788]: _type = "Task" [ 1013.335615] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.344901] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218ebae-7a18-8a87-8d3e-7902850d9b70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.401832] env[65788]: WARNING openstack [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.402445] env[65788]: WARNING openstack [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.424057] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.584s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.426940] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.124s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.427269] env[65788]: DEBUG nova.objects.instance [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'pci_requests' on Instance uuid a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.477950] env[65788]: DEBUG nova.compute.manager [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Received event network-vif-plugged-81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1013.478249] env[65788]: DEBUG oslo_concurrency.lockutils [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Acquiring lock "51705add-4d88-40bb-b236-2486650c77f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.478249] env[65788]: DEBUG oslo_concurrency.lockutils [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Lock "51705add-4d88-40bb-b236-2486650c77f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.478768] env[65788]: DEBUG oslo_concurrency.lockutils [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Lock "51705add-4d88-40bb-b236-2486650c77f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.478768] env[65788]: DEBUG nova.compute.manager [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] No waiting events found dispatching network-vif-plugged-81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1013.478768] env[65788]: WARNING nova.compute.manager [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Received unexpected event network-vif-plugged-81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 for instance with vm_state building and task_state spawning. [ 1013.478936] env[65788]: DEBUG nova.compute.manager [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Received event network-changed-81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1013.478988] env[65788]: DEBUG nova.compute.manager [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Refreshing instance network info cache due to event network-changed-81a17df2-1bdf-4672-a9d0-4fcd91b8edf6. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1013.479672] env[65788]: DEBUG oslo_concurrency.lockutils [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Acquiring lock "refresh_cache-51705add-4d88-40bb-b236-2486650c77f4" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.479672] env[65788]: DEBUG oslo_concurrency.lockutils [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Acquired lock "refresh_cache-51705add-4d88-40bb-b236-2486650c77f4" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.479672] env[65788]: DEBUG nova.network.neutron [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Refreshing network info cache for port 81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1013.507843] env[65788]: WARNING neutronclient.v2_0.client [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.508940] env[65788]: WARNING openstack [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.508940] env[65788]: WARNING openstack [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1013.565487] env[65788]: INFO nova.compute.manager [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Took 31.14 seconds to build instance. [ 1013.609201] env[65788]: DEBUG nova.network.neutron [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Updated VIF entry in instance network info cache for port bceffc39-3195-4064-9ae1-646b7a191ae7. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1013.609473] env[65788]: DEBUG nova.network.neutron [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Updating instance_info_cache with network_info: [{"id": "bceffc39-3195-4064-9ae1-646b7a191ae7", "address": "fa:16:3e:6a:ad:3e", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbceffc39-31", "ovs_interfaceid": "bceffc39-3195-4064-9ae1-646b7a191ae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1013.648553] env[65788]: INFO nova.compute.manager [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Shelve offloading [ 1013.787186] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "refresh_cache-51705add-4d88-40bb-b236-2486650c77f4" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.847672] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218ebae-7a18-8a87-8d3e-7902850d9b70, 'name': SearchDatastore_Task, 'duration_secs': 0.010496} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.848022] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1013.848473] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.848545] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.849091] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1013.849091] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.849537] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da6ded30-4ecb-435d-bca9-3b9a31cba835 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.859039] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.859327] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.860221] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f738656-2e49-4130-a290-4eedceafe569 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.866628] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1013.866628] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523891e6-bea8-e15b-d535-929b6989e174" [ 1013.866628] env[65788]: _type = "Task" [ 1013.866628] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.875203] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523891e6-bea8-e15b-d535-929b6989e174, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.933440] env[65788]: DEBUG nova.objects.instance [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'numa_topology' on Instance uuid a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.935611] env[65788]: DEBUG oslo_concurrency.lockutils [None req-68dc2441-af3c-4b62-b875-676859baee8f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 33.469s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.935874] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 12.085s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.939699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.939699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.939699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.939699] env[65788]: INFO nova.compute.manager [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Terminating instance [ 1013.981801] env[65788]: WARNING neutronclient.v2_0.client [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1013.982506] env[65788]: WARNING openstack [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1013.982871] env[65788]: WARNING openstack [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.025312] env[65788]: DEBUG nova.network.neutron [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1014.069661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-708b1c65-d78d-44fa-80bc-bb86c2e7cfed tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.658s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.113997] env[65788]: DEBUG oslo_concurrency.lockutils [req-50064b6b-f1bd-4282-88af-c05bd7fe332c req-3d40564d-30b8-4562-ae21-80c8d93d9ab4 service nova] Releasing lock "refresh_cache-6e78042d-ba10-4629-816f-3a13b2e22d4e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.126766] env[65788]: DEBUG nova.network.neutron [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1014.153500] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.153500] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52d9b293-255f-4cdd-b75c-d57e79c01264 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.162993] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1014.162993] env[65788]: value = "task-4662909" [ 1014.162993] env[65788]: _type = "Task" [ 1014.162993] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.176203] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1014.176445] env[65788]: DEBUG nova.compute.manager [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1014.179782] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7518b17-b4a2-4af7-8a36-5aa8d29fb2e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.188584] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.188891] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.189200] env[65788]: DEBUG nova.network.neutron [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1014.379341] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523891e6-bea8-e15b-d535-929b6989e174, 'name': SearchDatastore_Task, 'duration_secs': 0.010192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.379341] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-891efd41-4f27-44c5-8216-497fdc960bf5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.384887] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1014.384887] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52050a28-a8ca-ae95-604d-58f59364916e" [ 1014.384887] env[65788]: _type = "Task" [ 1014.384887] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.393435] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52050a28-a8ca-ae95-604d-58f59364916e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.436300] env[65788]: INFO nova.compute.claims [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.446494] env[65788]: DEBUG nova.compute.manager [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1014.446771] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1014.447114] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef680daa-6ff6-4aa6-8871-5ad1a16e2586 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.457673] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dffe3ae-96b8-44d9-8a80-4f3959b7cd75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.491789] env[65788]: WARNING nova.virt.vmwareapi.vmops [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 29c0ab17-5df2-4f6c-bdea-3a4a6c857066 could not be found. [ 1014.492015] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.492198] env[65788]: INFO nova.compute.manager [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1014.492459] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1014.493423] env[65788]: DEBUG nova.compute.manager [-] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1014.493608] env[65788]: DEBUG nova.network.neutron [-] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1014.493790] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.494396] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.494648] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.544204] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.630543] env[65788]: DEBUG oslo_concurrency.lockutils [req-2fdd0bb8-2b14-4cd4-a615-a7ff25c4548a req-cfe135ee-fe4c-4556-931e-5ecc70f7100d service nova] Releasing lock "refresh_cache-51705add-4d88-40bb-b236-2486650c77f4" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.631417] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "refresh_cache-51705add-4d88-40bb-b236-2486650c77f4" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1014.631651] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1014.684414] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.684579] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.684705] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.684848] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1014.685311] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.687515] env[65788]: INFO nova.compute.manager [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Terminating instance [ 1014.692472] env[65788]: WARNING neutronclient.v2_0.client [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.692677] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.693124] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.825928] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.826446] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.890755] env[65788]: WARNING neutronclient.v2_0.client [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1014.891501] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1014.891849] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1014.905779] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52050a28-a8ca-ae95-604d-58f59364916e, 'name': SearchDatastore_Task, 'duration_secs': 0.010231} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.906067] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1014.906296] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 6e78042d-ba10-4629-816f-3a13b2e22d4e/6e78042d-ba10-4629-816f-3a13b2e22d4e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.906567] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-209037f6-73be-4dba-b3ca-9b360a0f86d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.913394] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1014.913394] env[65788]: value = "task-4662911" [ 1014.913394] env[65788]: _type = "Task" [ 1014.913394] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.923707] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.981887] env[65788]: DEBUG nova.network.neutron [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.138578] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.139133] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.155435] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "902d8a69-0398-4752-a609-0ac5b4b32e27" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.155880] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.156139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "902d8a69-0398-4752-a609-0ac5b4b32e27-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.156358] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.156534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.159067] env[65788]: INFO nova.compute.manager [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Terminating instance [ 1015.188740] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1015.201528] env[65788]: DEBUG nova.compute.manager [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1015.201825] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.202848] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3931240b-7270-41f5-8526-e97e2dc02ee2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.210102] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.210541] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.220953] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.221277] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a790433-16a1-42ad-96dc-297707384374 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.230682] env[65788]: DEBUG oslo_vmware.api [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 1015.230682] env[65788]: value = "task-4662912" [ 1015.230682] env[65788]: _type = "Task" [ 1015.230682] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.236072] env[65788]: DEBUG nova.network.neutron [-] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.241471] env[65788]: DEBUG oslo_vmware.api [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.299287] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.301050] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.301308] env[65788]: WARNING openstack [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.424783] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509468} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.424911] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 6e78042d-ba10-4629-816f-3a13b2e22d4e/6e78042d-ba10-4629-816f-3a13b2e22d4e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.425140] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.425468] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-530b2be7-4a46-419a-bbdc-157af9e85a5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.434385] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1015.434385] env[65788]: value = "task-4662913" [ 1015.434385] env[65788]: _type = "Task" [ 1015.434385] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.448458] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.455640] env[65788]: DEBUG nova.network.neutron [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Updating instance_info_cache with network_info: [{"id": "81a17df2-1bdf-4672-a9d0-4fcd91b8edf6", "address": "fa:16:3e:8f:51:79", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a17df2-1b", "ovs_interfaceid": "81a17df2-1bdf-4672-a9d0-4fcd91b8edf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1015.487716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.488249] env[65788]: WARNING neutronclient.v2_0.client [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.488978] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1015.489475] env[65788]: WARNING openstack [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1015.494955] env[65788]: WARNING neutronclient.v2_0.client [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1015.666154] env[65788]: DEBUG nova.compute.manager [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1015.666154] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.667096] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ceb5c81-fed4-4792-acc7-46da1b614ad2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.676530] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.676838] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-647c326d-007b-4256-9ec7-5deabe1fc98c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.684119] env[65788]: DEBUG oslo_vmware.api [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1015.684119] env[65788]: value = "task-4662914" [ 1015.684119] env[65788]: _type = "Task" [ 1015.684119] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.698647] env[65788]: DEBUG oslo_vmware.api [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.741790] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a462834-d2a7-4b9f-8812-ee04b7e9e467 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.750755] env[65788]: INFO nova.compute.manager [-] [instance: 29c0ab17-5df2-4f6c-bdea-3a4a6c857066] Took 1.26 seconds to deallocate network for instance. [ 1015.751855] env[65788]: DEBUG oslo_vmware.api [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662912, 'name': PowerOffVM_Task, 'duration_secs': 0.325405} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.754871] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.755241] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.760083] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0390e389-d3af-4940-ba0b-706c665d3967 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.769820] env[65788]: DEBUG nova.compute.manager [req-a66903be-021e-4750-8040-2c208135f2a6 req-ffd76af5-7bf0-4fc1-8aa6-be79e7b22851 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-vif-unplugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1015.770143] env[65788]: DEBUG oslo_concurrency.lockutils [req-a66903be-021e-4750-8040-2c208135f2a6 req-ffd76af5-7bf0-4fc1-8aa6-be79e7b22851 service nova] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1015.770375] env[65788]: DEBUG oslo_concurrency.lockutils [req-a66903be-021e-4750-8040-2c208135f2a6 req-ffd76af5-7bf0-4fc1-8aa6-be79e7b22851 service nova] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1015.770544] env[65788]: DEBUG oslo_concurrency.lockutils [req-a66903be-021e-4750-8040-2c208135f2a6 req-ffd76af5-7bf0-4fc1-8aa6-be79e7b22851 service nova] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1015.770715] env[65788]: DEBUG nova.compute.manager [req-a66903be-021e-4750-8040-2c208135f2a6 req-ffd76af5-7bf0-4fc1-8aa6-be79e7b22851 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] No waiting events found dispatching network-vif-unplugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1015.770878] env[65788]: WARNING nova.compute.manager [req-a66903be-021e-4750-8040-2c208135f2a6 req-ffd76af5-7bf0-4fc1-8aa6-be79e7b22851 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received unexpected event network-vif-unplugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a for instance with vm_state shelved and task_state shelving_offloading. [ 1015.772931] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd02cd24-9168-4930-be6e-2b63f4ee7ae0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.808555] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf88cc43-8aac-43da-b18b-c1dd17401436 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.818409] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bacef71-a8a3-4f0b-9ff0-edcc85786281 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.834781] env[65788]: DEBUG nova.compute.provider_tree [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1015.869456] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.870402] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01812ff-e758-4464-87bf-58b68f7b5c61 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.879705] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.880026] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57e2865a-4791-4540-830f-d71299906ccc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.942400] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.942769] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.943503] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Deleting the datastore file [datastore2] 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.947255] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-842b281e-fecb-4115-bbcc-dd0f48d3c942 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.949442] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072985} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.949709] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.951035] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fddaf8e-9802-45d0-940f-5f011985ce20 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.955622] env[65788]: DEBUG oslo_vmware.api [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for the task: (returnval){ [ 1015.955622] env[65788]: value = "task-4662917" [ 1015.955622] env[65788]: _type = "Task" [ 1015.955622] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.970741] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "refresh_cache-51705add-4d88-40bb-b236-2486650c77f4" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1015.971215] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Instance network_info: |[{"id": "81a17df2-1bdf-4672-a9d0-4fcd91b8edf6", "address": "fa:16:3e:8f:51:79", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a17df2-1b", "ovs_interfaceid": "81a17df2-1bdf-4672-a9d0-4fcd91b8edf6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1015.980273] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 6e78042d-ba10-4629-816f-3a13b2e22d4e/6e78042d-ba10-4629-816f-3a13b2e22d4e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.982280] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:51:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81a17df2-1bdf-4672-a9d0-4fcd91b8edf6', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.989750] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1015.990813] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28d2e923-790e-4bbd-8728-b9db7d3f59bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.005667] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1016.005870] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1016.006062] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleting the datastore file [datastore1] 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.010730] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1016.010730] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f71b48de-dada-4344-8116-c59826792434 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.012565] env[65788]: DEBUG oslo_vmware.api [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.012822] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d341c350-b002-4eac-bf7d-fb8e3f5f5362 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.035105] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1016.035105] env[65788]: value = "task-4662918" [ 1016.035105] env[65788]: _type = "Task" [ 1016.035105] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.037648] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.037648] env[65788]: value = "task-4662920" [ 1016.037648] env[65788]: _type = "Task" [ 1016.037648] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.038182] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1016.038182] env[65788]: value = "task-4662919" [ 1016.038182] env[65788]: _type = "Task" [ 1016.038182] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.063539] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662920, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.063539] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662919, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.063539] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662918, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.195459] env[65788]: DEBUG oslo_vmware.api [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662914, 'name': PowerOffVM_Task, 'duration_secs': 0.475803} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.195746] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.195907] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1016.196200] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b0ecbf7-6304-4db4-bb7d-2e71b97b724b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.342841] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1016.342841] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1016.342841] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleting the datastore file [datastore2] 902d8a69-0398-4752-a609-0ac5b4b32e27 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.343327] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2d8dfc4-e6de-4572-b294-fa7377e4d0c7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.352667] env[65788]: DEBUG oslo_vmware.api [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for the task: (returnval){ [ 1016.352667] env[65788]: value = "task-4662923" [ 1016.352667] env[65788]: _type = "Task" [ 1016.352667] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.362959] env[65788]: DEBUG oslo_vmware.api [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.378039] env[65788]: DEBUG nova.scheduler.client.report [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1016.378757] env[65788]: DEBUG nova.compute.provider_tree [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 120 to 121 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1016.379056] env[65788]: DEBUG nova.compute.provider_tree [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.468013] env[65788]: DEBUG oslo_vmware.api [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Task: {'id': task-4662917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253028} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.468224] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.468433] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.468617] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.468791] env[65788]: INFO nova.compute.manager [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1016.469121] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1016.469345] env[65788]: DEBUG nova.compute.manager [-] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1016.469461] env[65788]: DEBUG nova.network.neutron [-] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1016.469812] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.470476] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.470826] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.510178] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.556300] env[65788]: DEBUG oslo_vmware.api [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4662918, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218998} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.563704] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.563957] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.564160] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.566417] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662920, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.566706] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662919, 'name': ReconfigVM_Task, 'duration_secs': 0.425421} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.567310] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 6e78042d-ba10-4629-816f-3a13b2e22d4e/6e78042d-ba10-4629-816f-3a13b2e22d4e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.568694] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24cd058d-38d7-4e02-b5c0-2e74ab72935d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.576925] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1016.576925] env[65788]: value = "task-4662924" [ 1016.576925] env[65788]: _type = "Task" [ 1016.576925] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.588761] env[65788]: INFO nova.scheduler.client.report [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted allocations for instance 4dcf8335-701d-4386-9de4-f14f5d375d1f [ 1016.598188] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662924, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.792274] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cd945682-4dc3-4373-a366-5528d6086f5f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "29c0ab17-5df2-4f6c-bdea-3a4a6c857066" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.856s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.863393] env[65788]: DEBUG oslo_vmware.api [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Task: {'id': task-4662923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199836} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.863777] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.863882] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.864134] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.864363] env[65788]: INFO nova.compute.manager [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1016.864659] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1016.864903] env[65788]: DEBUG nova.compute.manager [-] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1016.865041] env[65788]: DEBUG nova.network.neutron [-] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1016.865325] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.866027] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1016.866183] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1016.888011] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.460s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.888011] env[65788]: WARNING neutronclient.v2_0.client [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.889682] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.240s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.889976] env[65788]: DEBUG nova.objects.instance [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lazy-loading 'resources' on Instance uuid c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.912397] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1016.942222] env[65788]: INFO nova.network.neutron [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating port 2cd6eb89-f768-4ee6-93a2-386b83c42638 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1017.060569] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662920, 'name': CreateVM_Task, 'duration_secs': 0.664932} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.060805] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.062443] env[65788]: WARNING neutronclient.v2_0.client [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1017.062443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.062443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.062684] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1017.062926] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac6f0cd5-68db-4ac6-bc4d-f84f077db03e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.069416] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1017.069416] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527a7f70-4f53-85e4-6bc0-7ddfada75541" [ 1017.069416] env[65788]: _type = "Task" [ 1017.069416] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.078922] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527a7f70-4f53-85e4-6bc0-7ddfada75541, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.090378] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662924, 'name': Rename_Task, 'duration_secs': 0.187706} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.090644] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.091431] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98dc5969-e467-4050-8dee-420e3f43d39a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.100569] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1017.101010] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1017.101010] env[65788]: value = "task-4662925" [ 1017.101010] env[65788]: _type = "Task" [ 1017.101010] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.111299] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662925, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.502285] env[65788]: DEBUG nova.network.neutron [-] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.590033] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527a7f70-4f53-85e4-6bc0-7ddfada75541, 'name': SearchDatastore_Task, 'duration_secs': 0.015646} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.590033] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.590033] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1017.590033] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.590033] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1017.590033] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1017.592639] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a031355f-d294-46b8-b296-f5bb5056c072 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.607848] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1017.607848] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1017.616014] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1412b14-1d8e-44e0-a8bf-0d033e8f1c5f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.618615] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662925, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.624220] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1017.624220] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ffb8ce-b5ab-8896-1734-e328f35c8c3e" [ 1017.624220] env[65788]: _type = "Task" [ 1017.624220] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.637055] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ffb8ce-b5ab-8896-1734-e328f35c8c3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.712988] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfc8e93-daee-41c2-b6b8-f65560823a2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.721517] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367b1a26-63ca-4c90-b7a8-2447af65cf93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.759431] env[65788]: DEBUG nova.network.neutron [-] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1017.768110] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99a211d-9d10-434d-b86a-056f9fb25e04 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.776323] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-860ac531-19e7-4fce-9632-32550d566015 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.795035] env[65788]: DEBUG nova.compute.provider_tree [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.844370] env[65788]: DEBUG nova.compute.manager [req-ad842fb1-39cf-49d2-b0b3-6de724399076 req-7c42a7fe-f6f1-41fb-895e-9c572ccc99e8 service nova] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Received event network-vif-deleted-ec66b0ca-3c2b-4ee6-aa95-6b07e9e55c7b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1018.005643] env[65788]: INFO nova.compute.manager [-] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Took 1.54 seconds to deallocate network for instance. [ 1018.114761] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662925, 'name': PowerOnVM_Task, 'duration_secs': 0.67019} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.114949] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.115186] env[65788]: INFO nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Took 7.88 seconds to spawn the instance on the hypervisor. [ 1018.115435] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1018.116470] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ce9c6b-db07-4ea4-810b-3ed2f92e5ce0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.138010] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ffb8ce-b5ab-8896-1734-e328f35c8c3e, 'name': SearchDatastore_Task, 'duration_secs': 0.024427} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.139324] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14237488-c26d-4c88-9d14-422d9b892760 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.147273] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1018.147273] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528986c8-ba97-50cd-6d91-295917c04238" [ 1018.147273] env[65788]: _type = "Task" [ 1018.147273] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.153129] env[65788]: DEBUG nova.compute.manager [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1018.154239] env[65788]: DEBUG nova.compute.manager [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing instance network info cache due to event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1018.154239] env[65788]: DEBUG oslo_concurrency.lockutils [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.154239] env[65788]: DEBUG oslo_concurrency.lockutils [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.154239] env[65788]: DEBUG nova.network.neutron [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1018.162047] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528986c8-ba97-50cd-6d91-295917c04238, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.270654] env[65788]: INFO nova.compute.manager [-] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Took 1.41 seconds to deallocate network for instance. [ 1018.324414] env[65788]: ERROR nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] [req-fa468468-7fe0-43b9-8f5a-7b62078163cd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fa468468-7fe0-43b9-8f5a-7b62078163cd"}]} [ 1018.347637] env[65788]: DEBUG nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1018.369965] env[65788]: DEBUG nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1018.370632] env[65788]: DEBUG nova.compute.provider_tree [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.387537] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "64acef8e-640d-4195-9519-ce31d984ddde" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.387537] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.393481] env[65788]: DEBUG nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1018.417148] env[65788]: DEBUG nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1018.511870] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.642565] env[65788]: INFO nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Took 21.92 seconds to build instance. [ 1018.656454] env[65788]: WARNING neutronclient.v2_0.client [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.657167] env[65788]: WARNING openstack [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.657556] env[65788]: WARNING openstack [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.671815] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528986c8-ba97-50cd-6d91-295917c04238, 'name': SearchDatastore_Task, 'duration_secs': 0.030794} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.671815] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.672503] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 51705add-4d88-40bb-b236-2486650c77f4/51705add-4d88-40bb-b236-2486650c77f4.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1018.672503] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-818b0204-841c-4a9a-be50-7e244f2cce77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.681093] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1018.681093] env[65788]: value = "task-4662927" [ 1018.681093] env[65788]: _type = "Task" [ 1018.681093] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.694662] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.714304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.714304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.714483] env[65788]: DEBUG nova.network.neutron [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1018.731883] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e422061c-a84c-4bb9-9af5-0ade9197a3a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.743940] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d971abe-1d9d-4354-8065-e36ea8d18c66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.781107] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdd9574-7a12-494f-ae57-ccfdc30c39eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.784290] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.790329] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c47613-f28f-4339-9dd1-6b7ae651dde2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.806239] env[65788]: DEBUG nova.compute.provider_tree [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.831537] env[65788]: WARNING openstack [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.832025] env[65788]: WARNING openstack [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.891052] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1018.899835] env[65788]: WARNING neutronclient.v2_0.client [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1018.900651] env[65788]: WARNING openstack [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1018.901231] env[65788]: WARNING openstack [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1018.984715] env[65788]: DEBUG nova.network.neutron [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updated VIF entry in instance network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1018.985126] env[65788]: DEBUG nova.network.neutron [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.013557] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.145595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.435s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.195131] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662927, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.220378] env[65788]: WARNING neutronclient.v2_0.client [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.221740] env[65788]: WARNING openstack [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.222261] env[65788]: WARNING openstack [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.354209] env[65788]: DEBUG nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1019.354516] env[65788]: DEBUG nova.compute.provider_tree [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 123 to 124 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1019.354834] env[65788]: DEBUG nova.compute.provider_tree [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1019.430772] env[65788]: WARNING openstack [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.431351] env[65788]: WARNING openstack [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.439954] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.488722] env[65788]: DEBUG oslo_concurrency.lockutils [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.488914] env[65788]: DEBUG nova.compute.manager [req-383c2e91-027e-4d95-9532-fe991e51a6c9 req-8cf9dc83-b0ab-4056-813a-f16145b89c7f service nova] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Received event network-vif-deleted-79eedfab-2cdd-4443-821e-f2a69d774f98 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1019.506668] env[65788]: WARNING neutronclient.v2_0.client [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1019.507383] env[65788]: WARNING openstack [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1019.507773] env[65788]: WARNING openstack [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1019.594664] env[65788]: DEBUG nova.network.neutron [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1019.692748] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.834091} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.693081] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 51705add-4d88-40bb-b236-2486650c77f4/51705add-4d88-40bb-b236-2486650c77f4.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1019.693321] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.693599] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f07740cc-1178-462d-9011-b60962ae46a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.700652] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1019.700652] env[65788]: value = "task-4662928" [ 1019.700652] env[65788]: _type = "Task" [ 1019.700652] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.711588] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662928, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.860801] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.971s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.864085] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.945s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.864085] env[65788]: DEBUG nova.objects.instance [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'resources' on Instance uuid d78d36e6-a542-4ba8-9e29-b8a7b28c559e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.871668] env[65788]: DEBUG nova.compute.manager [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-vif-plugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1019.871904] env[65788]: DEBUG oslo_concurrency.lockutils [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.872154] env[65788]: DEBUG oslo_concurrency.lockutils [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1019.872299] env[65788]: DEBUG oslo_concurrency.lockutils [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1019.872456] env[65788]: DEBUG nova.compute.manager [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] No waiting events found dispatching network-vif-plugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1019.872610] env[65788]: WARNING nova.compute.manager [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received unexpected event network-vif-plugged-2cd6eb89-f768-4ee6-93a2-386b83c42638 for instance with vm_state shelved_offloaded and task_state spawning. [ 1019.872760] env[65788]: DEBUG nova.compute.manager [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-changed-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1019.872908] env[65788]: DEBUG nova.compute.manager [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Refreshing instance network info cache due to event network-changed-2cd6eb89-f768-4ee6-93a2-386b83c42638. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1019.873140] env[65788]: DEBUG oslo_concurrency.lockutils [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.881488] env[65788]: INFO nova.scheduler.client.report [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Deleted allocations for instance c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb [ 1020.096951] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.099538] env[65788]: DEBUG oslo_concurrency.lockutils [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.099727] env[65788]: DEBUG nova.network.neutron [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Refreshing network info cache for port 2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1020.212227] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662928, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073198} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.212227] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.212444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc51d1e5-4b11-4672-a1c0-f1038339ebb5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.236837] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 51705add-4d88-40bb-b236-2486650c77f4/51705add-4d88-40bb-b236-2486650c77f4.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.239239] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84040363-8b77-411b-ab12-a762fef024e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.260192] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7719f9c97da03bd96a7d11d292903d41',container_format='bare',created_at=2025-11-21T13:52:09Z,direct_url=,disk_format='vmdk',id=0f208f1a-1c47-4d89-ac7d-a4f584ec31ed,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-828800117-shelved',owner='5e7aa18dcc41403389518d700c2c8b8a',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2025-11-21T13:52:26Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1020.260507] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1020.260737] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1020.260930] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1020.261094] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1020.261248] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1020.261492] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.261660] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1020.261828] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1020.262035] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1020.262224] env[65788]: DEBUG nova.virt.hardware [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1020.263334] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50dc44f-5b35-444d-bdec-a561959d4cb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.268037] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1020.268037] env[65788]: value = "task-4662929" [ 1020.268037] env[65788]: _type = "Task" [ 1020.268037] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.275719] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874e5d59-7659-4ac6-8b12-85f77b8e09fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.284064] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662929, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.295975] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:d1:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cd6eb89-f768-4ee6-93a2-386b83c42638', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.303898] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1020.304197] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.304397] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2c95198-86d9-4216-831a-35febcda0804 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.324620] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.324620] env[65788]: value = "task-4662930" [ 1020.324620] env[65788]: _type = "Task" [ 1020.324620] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.334381] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662930, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.394463] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a28b262c-5618-403e-a756-0ddec8b433cb tempest-ImagesOneServerTestJSON-102617875 tempest-ImagesOneServerTestJSON-102617875-project-member] Lock "c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.670s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.608473] env[65788]: WARNING neutronclient.v2_0.client [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.609176] env[65788]: WARNING openstack [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.609530] env[65788]: WARNING openstack [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.726099] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a25877e-adf9-4fa0-925d-dd41d9239b06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.735218] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d4e5a1-f574-4531-b16e-056572b2a79f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.773175] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b661e921-bf03-473d-9383-43796de4a1b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.785346] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1ae98e-da04-4cae-a60c-26ae56d3e3ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.789360] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662929, 'name': ReconfigVM_Task, 'duration_secs': 0.437574} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.792045] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 51705add-4d88-40bb-b236-2486650c77f4/51705add-4d88-40bb-b236-2486650c77f4.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.793722] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af89e4e8-6f00-4e01-ad89-b7b9a2014e0f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.805068] env[65788]: DEBUG nova.compute.provider_tree [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.811209] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1020.811209] env[65788]: value = "task-4662932" [ 1020.811209] env[65788]: _type = "Task" [ 1020.811209] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.821955] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662932, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.836726] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662930, 'name': CreateVM_Task, 'duration_secs': 0.407927} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.836947] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.837557] env[65788]: WARNING neutronclient.v2_0.client [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1020.837936] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.838106] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.838479] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1020.839035] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed26ab50-071a-4feb-a503-9ed2f133e1ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.844340] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1020.844340] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525197f7-b7f0-7056-92d0-c550217cbcfc" [ 1020.844340] env[65788]: _type = "Task" [ 1020.844340] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.856183] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525197f7-b7f0-7056-92d0-c550217cbcfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.934850] env[65788]: WARNING openstack [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1020.935370] env[65788]: WARNING openstack [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1020.968789] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1020.969034] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.027968] env[65788]: WARNING neutronclient.v2_0.client [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1021.028883] env[65788]: WARNING openstack [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1021.029445] env[65788]: WARNING openstack [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1021.184670] env[65788]: DEBUG nova.network.neutron [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updated VIF entry in instance network info cache for port 2cd6eb89-f768-4ee6-93a2-386b83c42638. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1021.185077] env[65788]: DEBUG nova.network.neutron [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1021.313268] env[65788]: DEBUG nova.scheduler.client.report [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1021.328349] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662932, 'name': Rename_Task, 'duration_secs': 0.153896} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.329389] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.329490] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4dea2714-00a3-418b-bf18-194994ebb333 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.340264] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1021.340264] env[65788]: value = "task-4662933" [ 1021.340264] env[65788]: _type = "Task" [ 1021.340264] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.352734] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.360653] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.361157] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Processing image 0f208f1a-1c47-4d89-ac7d-a4f584ec31ed {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.361540] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.361803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1021.362127] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.362861] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-729ecab4-2951-44cb-ad55-9baf602af0cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.387252] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.387484] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.388349] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e4b1f3c-572e-46c7-affd-d55b4c5c19e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.395101] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1021.395101] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ba5193-9ef1-bef1-c335-ca52c12f5e2f" [ 1021.395101] env[65788]: _type = "Task" [ 1021.395101] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.405126] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ba5193-9ef1-bef1-c335-ca52c12f5e2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.476971] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.477350] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.477350] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.477509] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.477662] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.477809] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.477944] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1021.478465] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.688638] env[65788]: DEBUG oslo_concurrency.lockutils [req-aae82a34-20be-4463-9e62-0044dc16a104 req-cd90e071-d7bd-4100-b42b-d79b30fbe1a4 service nova] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1021.822536] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.824993] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.186s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.826996] env[65788]: INFO nova.compute.claims [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.854473] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662933, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.872510] env[65788]: INFO nova.scheduler.client.report [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleted allocations for instance d78d36e6-a542-4ba8-9e29-b8a7b28c559e [ 1021.906974] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1021.906974] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Fetch image to [datastore2] OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841/OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1021.906974] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Downloading stream optimized image 0f208f1a-1c47-4d89-ac7d-a4f584ec31ed to [datastore2] OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841/OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841.vmdk on the data store datastore2 as vApp {{(pid=65788) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1021.907292] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Downloading image file data 0f208f1a-1c47-4d89-ac7d-a4f584ec31ed to the ESX as VM named 'OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841' {{(pid=65788) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1021.982357] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1021.998326] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1021.998326] env[65788]: value = "resgroup-9" [ 1021.998326] env[65788]: _type = "ResourcePool" [ 1021.998326] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1021.998668] env[65788]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-550d0f1c-978b-4b73-884a-daa94da601be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.028069] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lease: (returnval){ [ 1022.028069] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c156ac-3838-f358-d9b3-51ec522cb607" [ 1022.028069] env[65788]: _type = "HttpNfcLease" [ 1022.028069] env[65788]: } obtained for vApp import into resource pool (val){ [ 1022.028069] env[65788]: value = "resgroup-9" [ 1022.028069] env[65788]: _type = "ResourcePool" [ 1022.028069] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1022.029773] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the lease: (returnval){ [ 1022.029773] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c156ac-3838-f358-d9b3-51ec522cb607" [ 1022.029773] env[65788]: _type = "HttpNfcLease" [ 1022.029773] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1022.039475] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1022.039475] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c156ac-3838-f358-d9b3-51ec522cb607" [ 1022.039475] env[65788]: _type = "HttpNfcLease" [ 1022.039475] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1022.350957] env[65788]: DEBUG oslo_vmware.api [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662933, 'name': PowerOnVM_Task, 'duration_secs': 0.806033} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.353378] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.353378] env[65788]: INFO nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Took 9.47 seconds to spawn the instance on the hypervisor. [ 1022.353378] env[65788]: DEBUG nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1022.354551] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159d48c4-7c70-41ce-b840-e23af2f2e569 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.385123] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2a40747f-5ba4-43db-bf7e-cd9629779c99 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "d78d36e6-a542-4ba8-9e29-b8a7b28c559e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.444s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.537480] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1022.537480] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c156ac-3838-f358-d9b3-51ec522cb607" [ 1022.537480] env[65788]: _type = "HttpNfcLease" [ 1022.537480] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1022.877560] env[65788]: INFO nova.compute.manager [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Took 26.08 seconds to build instance. [ 1022.918473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "9438ab56-1b4c-4778-a608-de319ab0ee43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1022.918473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.918473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "9438ab56-1b4c-4778-a608-de319ab0ee43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1022.918473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1022.918765] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1022.921291] env[65788]: INFO nova.compute.manager [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Terminating instance [ 1023.042286] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1023.042286] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c156ac-3838-f358-d9b3-51ec522cb607" [ 1023.042286] env[65788]: _type = "HttpNfcLease" [ 1023.042286] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1023.042286] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1023.042286] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c156ac-3838-f358-d9b3-51ec522cb607" [ 1023.042286] env[65788]: _type = "HttpNfcLease" [ 1023.042286] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1023.042286] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd21a179-46d9-482a-b4c9-1248c5af985d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.052571] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d82f62-08fe-efd5-a496-22f785c5fbc1/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1023.052779] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d82f62-08fe-efd5-a496-22f785c5fbc1/disk-0.vmdk. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1023.120595] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1a5e8939-8e2e-42ae-a17a-1c8497238267 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.189222] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526da139-0e09-4b47-b05e-080a917222cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.202492] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8119d8f-8fca-4fc9-be3f-b5494410f173 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.235769] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d59359-0c70-449f-9659-f565a8323623 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.245676] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4106687d-67e9-44f0-a2af-fc458e85d176 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.261825] env[65788]: DEBUG nova.compute.provider_tree [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.379833] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00548f4f-fc27-4cc1-917c-e36ac2ae577f tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "51705add-4d88-40bb-b236-2486650c77f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.604s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.425240] env[65788]: DEBUG nova.compute.manager [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1023.425489] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.426654] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4df5ed1-4ebc-4356-8055-43a7948eb390 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.437874] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.438179] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b147e16c-257e-4ee4-b8c9-459d771ed34d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.446939] env[65788]: DEBUG oslo_vmware.api [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 1023.446939] env[65788]: value = "task-4662936" [ 1023.446939] env[65788]: _type = "Task" [ 1023.446939] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.460087] env[65788]: DEBUG oslo_vmware.api [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662936, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.665641] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "6e78042d-ba10-4629-816f-3a13b2e22d4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.666036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.666180] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "6e78042d-ba10-4629-816f-3a13b2e22d4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.666365] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.666627] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.671482] env[65788]: INFO nova.compute.manager [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Terminating instance [ 1023.767648] env[65788]: DEBUG nova.scheduler.client.report [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1023.825344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "51705add-4d88-40bb-b236-2486650c77f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.825344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "51705add-4d88-40bb-b236-2486650c77f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.825344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "51705add-4d88-40bb-b236-2486650c77f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.825344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "51705add-4d88-40bb-b236-2486650c77f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.825344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "51705add-4d88-40bb-b236-2486650c77f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.828271] env[65788]: INFO nova.compute.manager [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Terminating instance [ 1023.962801] env[65788]: DEBUG oslo_vmware.api [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662936, 'name': PowerOffVM_Task, 'duration_secs': 0.345787} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.964728] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.964922] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.965623] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd191a9f-cc1e-4dc2-89c2-9f5190b0eb33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.988198] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.988441] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.046518] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1024.046966] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1024.047074] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleting the datastore file [datastore1] 9438ab56-1b4c-4778-a608-de319ab0ee43 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.051871] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89afed0b-fb21-4d91-b7ab-45fbe17e027d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.059733] env[65788]: DEBUG oslo_vmware.api [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for the task: (returnval){ [ 1024.059733] env[65788]: value = "task-4662938" [ 1024.059733] env[65788]: _type = "Task" [ 1024.059733] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.072377] env[65788]: DEBUG oslo_vmware.api [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662938, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.179288] env[65788]: DEBUG nova.compute.manager [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1024.179521] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.182017] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f288e9dd-7a4b-4ef8-8067-46c80afdc6b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.191509] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.191839] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f848dea-cddb-401a-acf0-0ed57786ddfa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.203030] env[65788]: DEBUG oslo_vmware.api [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1024.203030] env[65788]: value = "task-4662939" [ 1024.203030] env[65788]: _type = "Task" [ 1024.203030] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.214476] env[65788]: DEBUG oslo_vmware.api [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.276644] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.277231] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1024.280267] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 15.308s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1024.339377] env[65788]: DEBUG nova.compute.manager [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1024.339377] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1024.339377] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef09398e-4866-425d-bb7b-48ba17589d4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.347827] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.348261] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-240eafd3-5384-4ba9-8633-69028ef24a80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.360220] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1024.360220] env[65788]: value = "task-4662940" [ 1024.360220] env[65788]: _type = "Task" [ 1024.360220] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.373545] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.495864] env[65788]: DEBUG nova.compute.utils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1024.552082] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1024.552218] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d82f62-08fe-efd5-a496-22f785c5fbc1/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1024.554394] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb2f883-0b02-44ed-9c1b-bf832a3bf123 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.577699] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d82f62-08fe-efd5-a496-22f785c5fbc1/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1024.577699] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d82f62-08fe-efd5-a496-22f785c5fbc1/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1024.577699] env[65788]: DEBUG oslo_vmware.api [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Task: {'id': task-4662938, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274522} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.577995] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-bc81f3aa-c846-461f-bf6e-fdc658d6b9ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.580033] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.580189] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.580316] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.580592] env[65788]: INFO nova.compute.manager [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1024.580898] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1024.581092] env[65788]: DEBUG nova.compute.manager [-] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1024.581223] env[65788]: DEBUG nova.network.neutron [-] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1024.581510] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.582281] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.582377] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.637875] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.714189] env[65788]: DEBUG oslo_vmware.api [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662939, 'name': PowerOffVM_Task, 'duration_secs': 0.298371} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.714511] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1024.714749] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1024.714935] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76370a45-cb3a-4191-908b-f0c85bda2a08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.783956] env[65788]: DEBUG nova.compute.utils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1024.785601] env[65788]: DEBUG nova.objects.instance [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'migration_context' on Instance uuid 7f637326-9388-4d73-a1c7-3e4611ac46c5 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.786910] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1024.787146] env[65788]: DEBUG nova.network.neutron [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1024.790548] env[65788]: WARNING neutronclient.v2_0.client [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.790548] env[65788]: WARNING neutronclient.v2_0.client [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1024.790548] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1024.790548] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1024.803040] env[65788]: DEBUG oslo_vmware.rw_handles [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d82f62-08fe-efd5-a496-22f785c5fbc1/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1024.803040] env[65788]: INFO nova.virt.vmwareapi.images [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Downloaded image file data 0f208f1a-1c47-4d89-ac7d-a4f584ec31ed [ 1024.803312] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1024.803631] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1024.803717] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleting the datastore file [datastore2] 6e78042d-ba10-4629-816f-3a13b2e22d4e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.804733] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6fb67d-ee82-45c1-9204-564fdffaefe4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.807730] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22c38a1a-6ce7-4737-bad3-26222ba30311 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.825622] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5171686f-db09-449d-9084-3d25aa3ce660 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.827346] env[65788]: DEBUG oslo_vmware.api [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1024.827346] env[65788]: value = "task-4662942" [ 1024.827346] env[65788]: _type = "Task" [ 1024.827346] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.839758] env[65788]: DEBUG oslo_vmware.api [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.861747] env[65788]: INFO nova.virt.vmwareapi.images [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] The imported VM was unregistered [ 1024.864401] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1024.864648] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Creating directory with path [datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.865114] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32a05339-7f3d-44d6-baf1-fbdc45ce3574 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.879212] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662940, 'name': PowerOffVM_Task, 'duration_secs': 0.22754} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.881403] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1024.881504] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1024.886466] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f17669f-3ed8-4ada-8592-85b799856e83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.888427] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Created directory with path [datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.889167] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841/OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841.vmdk to [datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk. {{(pid=65788) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1024.889167] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-137db7cf-2fee-4239-bb6f-879c0ad5ed7f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.899968] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1024.899968] env[65788]: value = "task-4662945" [ 1024.899968] env[65788]: _type = "Task" [ 1024.899968] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.912040] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.966103] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1024.966484] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1024.967225] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleting the datastore file [datastore2] 51705add-4d88-40bb-b236-2486650c77f4 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.972526] env[65788]: DEBUG nova.policy [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b6bf354651e74188b7825c40e1840345', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '828bafa0635a42c7be55b1fb367e1734', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1024.975336] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a5719cf-db41-468a-a6d5-11dd3b7ec8f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.985220] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1024.985220] env[65788]: value = "task-4662946" [ 1024.985220] env[65788]: _type = "Task" [ 1024.985220] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.994710] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.999056] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.302957] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1025.345279] env[65788]: DEBUG oslo_vmware.api [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167148} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.346318] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.346513] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.346691] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.346861] env[65788]: INFO nova.compute.manager [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1025.347149] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1025.347460] env[65788]: DEBUG nova.compute.manager [-] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1025.347460] env[65788]: DEBUG nova.network.neutron [-] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1025.347687] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.348208] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1025.348490] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1025.386951] env[65788]: DEBUG nova.network.neutron [-] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1025.393196] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1025.399062] env[65788]: DEBUG nova.network.neutron [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Successfully created port: b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1025.418990] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.505046] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.661591] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097651ff-f63c-4a9c-8695-5d08a07cab92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.675471] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fe1dc7-dd12-46f5-ac47-6a77297c97e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.732557] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a90600-bb3f-4872-8ef5-65758e977d53 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.736568] env[65788]: DEBUG nova.compute.manager [req-26ee5829-9dbc-4fc5-92ca-00df024e4992 req-2745748b-7e6a-4496-ba6b-81ace25da2bf service nova] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Received event network-vif-deleted-f116c9d8-4c98-4af9-8747-0c585d99d738 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1025.747049] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cc0709-68fe-4a63-8818-fee779360922 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.764501] env[65788]: DEBUG nova.compute.provider_tree [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.890756] env[65788]: INFO nova.compute.manager [-] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Took 1.31 seconds to deallocate network for instance. [ 1025.914582] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.999022] env[65788]: DEBUG oslo_vmware.api [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4662946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.721417} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.999584] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.999584] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1025.999759] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1025.999943] env[65788]: INFO nova.compute.manager [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1026.000226] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1026.000437] env[65788]: DEBUG nova.compute.manager [-] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1026.000534] env[65788]: DEBUG nova.network.neutron [-] [instance: 51705add-4d88-40bb-b236-2486650c77f4] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1026.000910] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1026.001654] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1026.001766] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1026.091488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.091726] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.091961] env[65788]: INFO nova.compute.manager [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Attaching volume 676fc2ff-ea9f-4bd0-bd6a-86d399263acc to /dev/sdb [ 1026.130424] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ad8821-f886-4528-8707-64f24187d438 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.141039] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddefa763-7031-4993-8637-8996c4e98d5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.143765] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1026.161607] env[65788]: DEBUG nova.virt.block_device [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updating existing volume attachment record: 021ec389-9069-4413-b325-2c673dd77e79 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1026.187100] env[65788]: DEBUG nova.network.neutron [-] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1026.272114] env[65788]: DEBUG nova.scheduler.client.report [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1026.316229] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1026.358025] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1026.358495] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1026.358495] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1026.358691] env[65788]: DEBUG nova.virt.hardware [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1026.359819] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cd5f38-f360-407a-a19c-afde512558a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.371521] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df533cd5-05e9-4f6d-a74d-a0f1b722d360 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.399683] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.416813] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.691638] env[65788]: INFO nova.compute.manager [-] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Took 1.34 seconds to deallocate network for instance. [ 1026.917950] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.996590] env[65788]: DEBUG nova.network.neutron [-] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1027.185580] env[65788]: DEBUG nova.network.neutron [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Successfully updated port: b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1027.199386] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.283172] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.003s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.291224] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.527s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.291548] env[65788]: DEBUG nova.objects.instance [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lazy-loading 'resources' on Instance uuid 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.313856] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.314142] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.415596] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.469274] env[65788]: DEBUG nova.compute.manager [req-049e9404-0a6f-46e1-9c9e-0dd3b49fe4c7 req-696d9bb0-33dc-4a88-a558-ec87a1f746c7 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Received event network-vif-plugged-b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1027.469469] env[65788]: DEBUG oslo_concurrency.lockutils [req-049e9404-0a6f-46e1-9c9e-0dd3b49fe4c7 req-696d9bb0-33dc-4a88-a558-ec87a1f746c7 service nova] Acquiring lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.469662] env[65788]: DEBUG oslo_concurrency.lockutils [req-049e9404-0a6f-46e1-9c9e-0dd3b49fe4c7 req-696d9bb0-33dc-4a88-a558-ec87a1f746c7 service nova] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.469822] env[65788]: DEBUG oslo_concurrency.lockutils [req-049e9404-0a6f-46e1-9c9e-0dd3b49fe4c7 req-696d9bb0-33dc-4a88-a558-ec87a1f746c7 service nova] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.470013] env[65788]: DEBUG nova.compute.manager [req-049e9404-0a6f-46e1-9c9e-0dd3b49fe4c7 req-696d9bb0-33dc-4a88-a558-ec87a1f746c7 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] No waiting events found dispatching network-vif-plugged-b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1027.470200] env[65788]: WARNING nova.compute.manager [req-049e9404-0a6f-46e1-9c9e-0dd3b49fe4c7 req-696d9bb0-33dc-4a88-a558-ec87a1f746c7 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Received unexpected event network-vif-plugged-b382ad0a-1833-409d-8d27-6726195bc924 for instance with vm_state building and task_state spawning. [ 1027.499320] env[65788]: INFO nova.compute.manager [-] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Took 1.50 seconds to deallocate network for instance. [ 1027.554528] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "18be7322-3359-49c9-a181-f9228d5c16d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.554715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.690045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.690892] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquired lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.690892] env[65788]: DEBUG nova.network.neutron [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1027.757341] env[65788]: DEBUG nova.compute.manager [req-bda5b711-2772-43e5-8792-6af420ec6c94 req-8dcb6321-9f0b-4a7e-8da6-b766ff46a022 service nova] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Received event network-vif-deleted-bceffc39-3195-4064-9ae1-646b7a191ae7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1027.757574] env[65788]: DEBUG nova.compute.manager [req-bda5b711-2772-43e5-8792-6af420ec6c94 req-8dcb6321-9f0b-4a7e-8da6-b766ff46a022 service nova] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Received event network-vif-deleted-81a17df2-1bdf-4672-a9d0-4fcd91b8edf6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1027.816460] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1027.917227] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662945, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.596621} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.920101] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841/OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841.vmdk to [datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk. [ 1027.920309] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Cleaning up location [datastore2] OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841 {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1027.920469] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_508aeda8-096d-4c1a-bc3c-341d50d07841 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.920989] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cde280d3-326c-4d57-b1ac-c7320e5dec5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.929383] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1027.929383] env[65788]: value = "task-4662950" [ 1027.929383] env[65788]: _type = "Task" [ 1027.929383] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.944457] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662950, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.010771] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.056525] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1028.141111] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a68619d-c5e0-4316-b176-bec5a5a235ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.151657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bed176-6206-44ba-abc7-116a26142824 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.185526] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0715763f-9bd1-4e78-9521-04ccfdf39578 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.197126] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.197126] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.206953] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015e8027-da13-40a8-b4a8-b86f07ea7f26 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.222082] env[65788]: DEBUG nova.compute.provider_tree [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.247491] env[65788]: DEBUG nova.network.neutron [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1028.266032] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.266472] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.337069] env[65788]: WARNING neutronclient.v2_0.client [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.337069] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1028.337069] env[65788]: WARNING openstack [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1028.345941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.438954] env[65788]: DEBUG nova.network.neutron [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Updating instance_info_cache with network_info: [{"id": "b382ad0a-1833-409d-8d27-6726195bc924", "address": "fa:16:3e:a2:c3:61", "network": {"id": "434cda43-035e-4e61-a5b8-0145c358571f", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1972275156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828bafa0635a42c7be55b1fb367e1734", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb382ad0a-18", "ovs_interfaceid": "b382ad0a-1833-409d-8d27-6726195bc924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1028.443470] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662950, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03959} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.444023] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1028.444239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.444485] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk to [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.444757] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15e9d609-8937-4886-bcc4-665920102571 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.459685] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1028.459685] env[65788]: value = "task-4662951" [ 1028.459685] env[65788]: _type = "Task" [ 1028.459685] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.469919] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.579214] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1028.726067] env[65788]: DEBUG nova.scheduler.client.report [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.829823] env[65788]: INFO nova.compute.manager [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Swapping old allocation on dict_keys(['3a6417f0-b161-4086-8a0e-1cb139eae377']) held by migration 761247b6-8ee4-4c67-b8fb-cfd2656b3d8d for instance [ 1028.858188] env[65788]: DEBUG nova.scheduler.client.report [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Overwriting current allocation {'allocations': {'3a6417f0-b161-4086-8a0e-1cb139eae377': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 126}}, 'project_id': '2b70d3c8627449eaa6372ebe3bd90233', 'user_id': '2625758e73c64384982cb820ea055cb1', 'consumer_generation': 1} on consumer 7f637326-9388-4d73-a1c7-3e4611ac46c5 {{(pid=65788) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1028.913030] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1028.944473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.944682] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.944887] env[65788]: DEBUG nova.network.neutron [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1028.946565] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Releasing lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.946880] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Instance network_info: |[{"id": "b382ad0a-1833-409d-8d27-6726195bc924", "address": "fa:16:3e:a2:c3:61", "network": {"id": "434cda43-035e-4e61-a5b8-0145c358571f", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1972275156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828bafa0635a42c7be55b1fb367e1734", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb382ad0a-18", "ovs_interfaceid": "b382ad0a-1833-409d-8d27-6726195bc924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1028.947734] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:c3:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b5c60ce-845e-4506-bc10-348461fece6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b382ad0a-1833-409d-8d27-6726195bc924', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.955798] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Creating folder: Project (828bafa0635a42c7be55b1fb367e1734). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.956933] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b1d2dac-f705-4b27-90ac-9af306f16a27 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.970237] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662951, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.972063] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Created folder: Project (828bafa0635a42c7be55b1fb367e1734) in parent group-v910111. [ 1028.972235] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Creating folder: Instances. Parent ref: group-v910385. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1028.972469] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3688f03d-3619-424a-bd20-c551c16d8b4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.987023] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Created folder: Instances in parent group-v910385. [ 1028.987353] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1028.987573] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.987828] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c09c2cff-d51c-4fff-b4b9-a193fefb11cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.013092] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.013092] env[65788]: value = "task-4662955" [ 1029.013092] env[65788]: _type = "Task" [ 1029.013092] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.024259] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662955, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.235036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.942s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.235595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.998s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1029.235848] env[65788]: DEBUG nova.objects.instance [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lazy-loading 'resources' on Instance uuid 2bcfbe11-51a2-49b4-b482-02ca332d8c38 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.257151] env[65788]: INFO nova.scheduler.client.report [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Deleted allocations for instance 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf [ 1029.448414] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.449169] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.449587] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.471962] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662951, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.527504] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662955, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.551264] env[65788]: DEBUG nova.compute.manager [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Received event network-changed-b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1029.551264] env[65788]: DEBUG nova.compute.manager [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Refreshing instance network info cache due to event network-changed-b382ad0a-1833-409d-8d27-6726195bc924. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1029.551457] env[65788]: DEBUG oslo_concurrency.lockutils [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Acquiring lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.551457] env[65788]: DEBUG oslo_concurrency.lockutils [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Acquired lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.551589] env[65788]: DEBUG nova.network.neutron [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Refreshing network info cache for port b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1029.604024] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.604024] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.668204] env[65788]: WARNING neutronclient.v2_0.client [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1029.669029] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1029.669406] env[65788]: WARNING openstack [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1029.766541] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12e83864-8f50-4ef5-97dd-51601eaf6d3e tempest-ServersTestMultiNic-291669374 tempest-ServersTestMultiNic-291669374-project-member] Lock "4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.931s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1029.782142] env[65788]: DEBUG nova.network.neutron [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [{"id": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "address": "fa:16:3e:3e:9d:19", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape98f1b15-0a", "ovs_interfaceid": "e98f1b15-0a05-4be4-b950-dd7e6749eb8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1029.976430] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662951, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.019397] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4bad11-2160-40c3-881c-031951c02ce2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.032099] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a0447d-fdac-4400-b241-6d6cf6f83229 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.035684] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662955, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.069031] env[65788]: WARNING neutronclient.v2_0.client [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.069670] env[65788]: WARNING openstack [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.070385] env[65788]: WARNING openstack [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.080742] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9450f0e-a65a-4050-a121-7065926467ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.092198] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1eb08b6-19cb-47cc-a9c4-8cb1e1610536 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.112445] env[65788]: DEBUG nova.compute.provider_tree [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.200428] env[65788]: WARNING openstack [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.200861] env[65788]: WARNING openstack [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.269686] env[65788]: WARNING neutronclient.v2_0.client [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.271143] env[65788]: WARNING openstack [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1030.271794] env[65788]: WARNING openstack [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1030.285079] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-7f637326-9388-4d73-a1c7-3e4611ac46c5" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.285997] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31dc012c-ff9b-47be-8bd4-04741338d172 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.297070] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4fa755-a148-496b-adee-4e445123366c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.380173] env[65788]: DEBUG nova.network.neutron [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Updated VIF entry in instance network info cache for port b382ad0a-1833-409d-8d27-6726195bc924. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1030.380173] env[65788]: DEBUG nova.network.neutron [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Updating instance_info_cache with network_info: [{"id": "b382ad0a-1833-409d-8d27-6726195bc924", "address": "fa:16:3e:a2:c3:61", "network": {"id": "434cda43-035e-4e61-a5b8-0145c358571f", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1972275156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828bafa0635a42c7be55b1fb367e1734", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb382ad0a-18", "ovs_interfaceid": "b382ad0a-1833-409d-8d27-6726195bc924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1030.480127] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662951, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.529465] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662955, 'name': CreateVM_Task, 'duration_secs': 1.066506} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.529465] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.530030] env[65788]: WARNING neutronclient.v2_0.client [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1030.530462] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.530704] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1030.531137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1030.532399] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aa98268-7a34-477f-8ffb-aebfae7039ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.540480] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1030.540480] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b3048a-e62f-ebac-4595-132274ac45e9" [ 1030.540480] env[65788]: _type = "Task" [ 1030.540480] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.552835] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b3048a-e62f-ebac-4595-132274ac45e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.617451] env[65788]: DEBUG nova.scheduler.client.report [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1030.724941] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1030.725297] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910384', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'name': 'volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ec7f7a46-8dd2-4b75-866f-20e73907f1cd', 'attached_at': '', 'detached_at': '', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'serial': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1030.726291] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193069f4-ff82-433a-ade0-f4fd44117fd8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.747161] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4529dd22-4f3f-430a-ad1d-c529d3efc277 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.776569] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc/volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.777383] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88297f4e-0a65-4fcb-b63e-712d41a075bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.799986] env[65788]: DEBUG oslo_vmware.api [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1030.799986] env[65788]: value = "task-4662956" [ 1030.799986] env[65788]: _type = "Task" [ 1030.799986] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.812682] env[65788]: DEBUG oslo_vmware.api [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662956, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.883184] env[65788]: DEBUG oslo_concurrency.lockutils [req-6ad96143-e47d-46d2-a2fd-0d2f00f27445 req-5c85339c-2abf-4b7b-ad43-29c2f62c8a14 service nova] Releasing lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.974567] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662951, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.479236} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.974792] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed/0f208f1a-1c47-4d89-ac7d-a4f584ec31ed.vmdk to [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1030.975596] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eef284a-6674-4779-b519-c3232cd62f45 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.998507] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.998847] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e587e268-5bbe-4356-98dc-7d8d55a5ce6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.020935] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1031.020935] env[65788]: value = "task-4662957" [ 1031.020935] env[65788]: _type = "Task" [ 1031.020935] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.032089] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662957, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.053582] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b3048a-e62f-ebac-4595-132274ac45e9, 'name': SearchDatastore_Task, 'duration_secs': 0.09289} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.053935] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.054195] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.054477] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.054723] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1031.055018] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.055386] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84a4cd7d-e5ee-45c2-816c-04f118df61f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.066054] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.066265] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.067054] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0b0fcfc-ac89-4943-893d-5bdbf9acb718 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.073322] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1031.073322] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5272fe14-2304-13b8-646b-bbcb93cdaaef" [ 1031.073322] env[65788]: _type = "Task" [ 1031.073322] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.086987] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5272fe14-2304-13b8-646b-bbcb93cdaaef, 'name': SearchDatastore_Task, 'duration_secs': 0.010622} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.088048] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9874f85e-39cc-46d7-8d7d-f368557da54d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.095053] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1031.095053] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e289aa-41ff-00f0-0116-7aba4ac3f3ab" [ 1031.095053] env[65788]: _type = "Task" [ 1031.095053] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.106802] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e289aa-41ff-00f0-0116-7aba4ac3f3ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.122540] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.886s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.124975] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.115s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.125452] env[65788]: DEBUG nova.objects.instance [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lazy-loading 'resources' on Instance uuid 374564df-70fa-4b89-8f38-e559245e5ebf {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.154483] env[65788]: INFO nova.scheduler.client.report [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Deleted allocations for instance 2bcfbe11-51a2-49b4-b482-02ca332d8c38 [ 1031.310163] env[65788]: DEBUG oslo_vmware.api [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662956, 'name': ReconfigVM_Task, 'duration_secs': 0.407319} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.310522] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Reconfigured VM instance instance-0000004d to attach disk [datastore2] volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc/volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.315638] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4913fa9-07ae-4310-b79f-03655ad94186 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.336816] env[65788]: DEBUG oslo_vmware.api [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1031.336816] env[65788]: value = "task-4662958" [ 1031.336816] env[65788]: _type = "Task" [ 1031.336816] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.352796] env[65788]: DEBUG oslo_vmware.api [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662958, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.428945] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.429288] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca124cc3-ba73-48dc-ad75-ecfc03dd3a37 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.437860] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1031.437860] env[65788]: value = "task-4662959" [ 1031.437860] env[65788]: _type = "Task" [ 1031.437860] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.449651] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662959, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.532405] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662957, 'name': ReconfigVM_Task, 'duration_secs': 0.334718} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.532851] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Reconfigured VM instance instance-0000003a to attach disk [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64/a35ced42-4317-49b4-b4cc-4ed7e2c85c64.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1031.533733] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62413674-1e45-4fd7-8dcf-c3de48297da2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.542885] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1031.542885] env[65788]: value = "task-4662960" [ 1031.542885] env[65788]: _type = "Task" [ 1031.542885] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.555345] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662960, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.605931] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e289aa-41ff-00f0-0116-7aba4ac3f3ab, 'name': SearchDatastore_Task, 'duration_secs': 0.013951} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.606279] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1031.606546] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 9d2f0d5e-793d-4790-ad31-be3d372be4fa/9d2f0d5e-793d-4790-ad31-be3d372be4fa.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.606825] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-676fa6e9-0e96-4e43-96e5-2ee08ad8d4a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.617339] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1031.617339] env[65788]: value = "task-4662961" [ 1031.617339] env[65788]: _type = "Task" [ 1031.617339] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.626884] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.666813] env[65788]: DEBUG oslo_concurrency.lockutils [None req-048c6c4a-ba41-46fb-8c1e-15a2780dec26 tempest-ImagesOneServerNegativeTestJSON-1524253690 tempest-ImagesOneServerNegativeTestJSON-1524253690-project-member] Lock "2bcfbe11-51a2-49b4-b482-02ca332d8c38" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.971s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.848744] env[65788]: DEBUG oslo_vmware.api [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4662958, 'name': ReconfigVM_Task, 'duration_secs': 0.155082} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.849540] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910384', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'name': 'volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ec7f7a46-8dd2-4b75-866f-20e73907f1cd', 'attached_at': '', 'detached_at': '', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'serial': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1031.949781] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662959, 'name': PowerOffVM_Task, 'duration_secs': 0.24854} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.951111] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1031.951822] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1031.952102] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1031.952277] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1031.952469] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1031.952609] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1031.952751] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1031.952962] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.953134] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1031.953426] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1031.953598] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1031.953806] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1031.963909] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfcfcb66-86dd-4632-9c1b-c27d24d9427e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.985730] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1031.985730] env[65788]: value = "task-4662962" [ 1031.985730] env[65788]: _type = "Task" [ 1031.985730] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.002422] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662962, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.062964] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662960, 'name': Rename_Task, 'duration_secs': 0.163342} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.063370] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1032.063649] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-597aa627-562f-4e1a-b845-1562e8ce097c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.077619] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1032.077619] env[65788]: value = "task-4662963" [ 1032.077619] env[65788]: _type = "Task" [ 1032.077619] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.086627] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e050ffa3-3185-43bb-9cb0-7a559143b1ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.093483] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.102134] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbe18d8-0d40-425f-81b4-99ea7b54d24f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.147727] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd4ac2c-8ed8-4046-8e6d-23c7448eefac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.158257] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662961, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.162313] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09e96a9-1cdc-453b-a369-e3efdf1bee15 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.184329] env[65788]: DEBUG nova.compute.provider_tree [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.498506] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662962, 'name': ReconfigVM_Task, 'duration_secs': 0.319801} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.500450] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcde50f-a809-45c2-bf5d-2a012383ebeb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.529135] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1032.529406] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.529562] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1032.529751] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.529878] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1032.530038] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1032.530232] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1032.530403] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1032.531512] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1032.531512] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1032.531512] env[65788]: DEBUG nova.virt.hardware [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1032.531903] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e583be64-f242-4725-a4da-46433d4cdd72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.539526] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1032.539526] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52735ad8-5194-9083-5a3f-9a5052fcb620" [ 1032.539526] env[65788]: _type = "Task" [ 1032.539526] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.550608] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52735ad8-5194-9083-5a3f-9a5052fcb620, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.588305] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662963, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.657788] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662961, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.583523} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.658098] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 9d2f0d5e-793d-4790-ad31-be3d372be4fa/9d2f0d5e-793d-4790-ad31-be3d372be4fa.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.658435] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.659063] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e558063-d52e-4aaa-ac40-aa813804815f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.667900] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1032.667900] env[65788]: value = "task-4662964" [ 1032.667900] env[65788]: _type = "Task" [ 1032.667900] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.683723] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.686687] env[65788]: DEBUG nova.scheduler.client.report [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1032.901568] env[65788]: DEBUG nova.objects.instance [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'flavor' on Instance uuid ec7f7a46-8dd2-4b75-866f-20e73907f1cd {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.053660] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52735ad8-5194-9083-5a3f-9a5052fcb620, 'name': SearchDatastore_Task, 'duration_secs': 0.011375} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.059546] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1033.062201] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a029ea77-a976-4d47-98d6-fe003498e6c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.085668] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1033.085668] env[65788]: value = "task-4662965" [ 1033.085668] env[65788]: _type = "Task" [ 1033.085668] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.097593] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662963, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.107883] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662965, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.179924] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.278474} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.180645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.181907] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad6e4ad-b0cf-44d8-9100-7c4e9236a67e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.198322] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.073s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.211833] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 9d2f0d5e-793d-4790-ad31-be3d372be4fa/9d2f0d5e-793d-4790-ad31-be3d372be4fa.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.211833] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.110s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.211833] env[65788]: DEBUG nova.objects.instance [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'resources' on Instance uuid 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.213306] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bf42880-0241-429a-ba85-51160158d5e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.236878] env[65788]: DEBUG nova.objects.instance [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'numa_topology' on Instance uuid 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.243306] env[65788]: INFO nova.scheduler.client.report [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Deleted allocations for instance 374564df-70fa-4b89-8f38-e559245e5ebf [ 1033.247303] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1033.247303] env[65788]: value = "task-4662966" [ 1033.247303] env[65788]: _type = "Task" [ 1033.247303] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.263083] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662966, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.410789] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fbd0f967-512f-4722-a6f6-3fe3c9878347 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.319s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.592814] env[65788]: DEBUG oslo_vmware.api [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662963, 'name': PowerOnVM_Task, 'duration_secs': 1.4251} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.598638] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1033.612932] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662965, 'name': ReconfigVM_Task, 'duration_secs': 0.379245} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.612932] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1033.613865] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4c714a-382a-43cf-a0c4-97c06df21963 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.648395] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.648783] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-733a9256-05f6-4d32-9adf-85afcae02bba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.671420] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1033.671420] env[65788]: value = "task-4662967" [ 1033.671420] env[65788]: _type = "Task" [ 1033.671420] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.683601] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662967, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.741457] env[65788]: DEBUG nova.objects.base [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Object Instance<4dcf8335-701d-4386-9de4-f14f5d375d1f> lazy-loaded attributes: resources,numa_topology {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1033.758234] env[65788]: DEBUG nova.compute.manager [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1033.759830] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047e275e-d0f6-48f9-bfbb-1f9878e6c2a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.763520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3209e8e-932d-4370-92ad-01d9f606736a tempest-SecurityGroupsTestJSON-451462583 tempest-SecurityGroupsTestJSON-451462583-project-member] Lock "374564df-70fa-4b89-8f38-e559245e5ebf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.235s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.768328] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662966, 'name': ReconfigVM_Task, 'duration_secs': 0.388612} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.769706] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 9d2f0d5e-793d-4790-ad31-be3d372be4fa/9d2f0d5e-793d-4790-ad31-be3d372be4fa.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.770864] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d478cd7-b6c9-4948-85a6-f9081e100bdf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.791035] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1033.791035] env[65788]: value = "task-4662968" [ 1033.791035] env[65788]: _type = "Task" [ 1033.791035] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.804521] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662968, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.074352] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52364cfd-197f-4249-ae94-fd6ff92ce063 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.087112] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b37f9b3-8804-49d5-8cfb-18f0867c3d4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.130036] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94918249-97e6-43f8-9b33-79e8c89c9e38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.142690] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595eb8f5-465f-4e24-a1f4-104773104458 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.165737] env[65788]: DEBUG nova.compute.provider_tree [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.182503] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.300798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c579bf94-4284-4fa8-bd0f-0ff62534fa79 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 30.030s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1034.310010] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662968, 'name': Rename_Task, 'duration_secs': 0.183085} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.310403] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.310752] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d03c03a7-d058-49f7-aa03-ab2706f5213b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.321134] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1034.321134] env[65788]: value = "task-4662969" [ 1034.321134] env[65788]: _type = "Task" [ 1034.321134] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.333281] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.673427] env[65788]: DEBUG nova.scheduler.client.report [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1034.689311] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662967, 'name': ReconfigVM_Task, 'duration_secs': 0.666275} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.689311] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5/7f637326-9388-4d73-a1c7-3e4611ac46c5.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.689805] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8aab7f-4cd3-4da1-85c2-d86e5d295ab8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.719190] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c68a6b-b659-4153-9f70-a68a606bb2bc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.746270] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fa7e90-95ce-4a41-a7ab-844f820f45ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.773243] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f41841-41b6-42b1-b592-32ac3fce7d85 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.783490] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.783925] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe9cd7cf-008a-452d-a4a7-65cbdd1c61d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.793855] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1034.793855] env[65788]: value = "task-4662970" [ 1034.793855] env[65788]: _type = "Task" [ 1034.793855] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.806586] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.832015] env[65788]: DEBUG oslo_vmware.api [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4662969, 'name': PowerOnVM_Task, 'duration_secs': 0.503017} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.832381] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.832530] env[65788]: INFO nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Took 8.52 seconds to spawn the instance on the hypervisor. [ 1034.832707] env[65788]: DEBUG nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1034.834097] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd260a2e-a08d-406d-b10f-98b1051bfa22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.184629] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.189511] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.678s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.190068] env[65788]: DEBUG nova.objects.instance [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lazy-loading 'resources' on Instance uuid 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.306760] env[65788]: DEBUG oslo_vmware.api [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662970, 'name': PowerOnVM_Task, 'duration_secs': 0.410377} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.307288] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.352990] env[65788]: INFO nova.compute.manager [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Took 26.74 seconds to build instance. [ 1035.699557] env[65788]: DEBUG oslo_concurrency.lockutils [None req-084374a7-f89d-4b17-b745-2d481a4cb189 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 40.170s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.700462] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 16.687s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.700734] env[65788]: INFO nova.compute.manager [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Unshelving [ 1035.854909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-48690ab9-1989-48a3-a76a-7e97931fe0f5 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.251s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.948307] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3948f29-1e2a-4d1b-93bc-06bd5cd63aae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.959511] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f253d161-25b6-4911-8e8b-6b6187f40ba3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.995807] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8c09d3-eaa4-4b18-bd36-feb2eebcb5a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.003019] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe55dc1-f2e7-400e-81c4-027e4f059613 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.022953] env[65788]: DEBUG nova.compute.provider_tree [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.334356] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.335284] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.372857] env[65788]: INFO nova.compute.manager [None req-3a486515-dac6-4c0d-b01b-820352ac7ab6 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance to original state: 'active' [ 1036.526579] env[65788]: DEBUG nova.scheduler.client.report [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1036.714986] env[65788]: DEBUG nova.compute.utils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1036.837943] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1037.033236] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.036419] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.252s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1037.036805] env[65788]: DEBUG nova.objects.instance [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lazy-loading 'resources' on Instance uuid 902d8a69-0398-4752-a609-0ac5b4b32e27 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.070093] env[65788]: INFO nova.scheduler.client.report [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Deleted allocations for instance 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc [ 1037.196672] env[65788]: DEBUG nova.compute.manager [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Received event network-changed-b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1037.196732] env[65788]: DEBUG nova.compute.manager [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Refreshing instance network info cache due to event network-changed-b382ad0a-1833-409d-8d27-6726195bc924. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1037.197535] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Acquiring lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.197535] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Acquired lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1037.197535] env[65788]: DEBUG nova.network.neutron [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Refreshing network info cache for port b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1037.224060] env[65788]: INFO nova.virt.block_device [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Booting with volume 7fc1ca8c-4392-42f3-9430-fec6de904bb6 at /dev/sdb [ 1037.271807] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff68f1ac-9bd6-4585-a97c-7d4a06763a63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.282997] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70048e7-cc2d-4d6b-b43e-c265e255c0d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.330118] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c572ec57-3539-4230-8aa6-e5a14f0db737 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.340911] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4055df10-7550-406f-9119-bf49e1958c10 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.380933] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105c6286-5047-432f-b44c-b83ef68d1e22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.385378] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1037.394851] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcb9670-cdbc-4438-8c9a-6ad381050d91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.414966] env[65788]: DEBUG nova.virt.block_device [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating existing volume attachment record: 6d58f422-5a00-4162-b476-54f8f15b4bc8 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1037.583573] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b68f5835-1cc8-48b7-ac56-32f7652182ec tempest-ServersTestManualDisk-213355435 tempest-ServersTestManualDisk-213355435-project-member] Lock "22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.899s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.703026] env[65788]: WARNING neutronclient.v2_0.client [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1037.703904] env[65788]: WARNING openstack [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.704359] env[65788]: WARNING openstack [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.850617] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b9884f-5c24-4703-a67a-ca1c6096edb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.859319] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb37f21-b81e-44ac-b6ce-2d2a707e8cf8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.891746] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df1a266-fb3e-4174-926c-bd2c7d609346 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.902134] env[65788]: WARNING openstack [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1037.902535] env[65788]: WARNING openstack [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1037.910417] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63892f79-93eb-411a-b134-225af870dbda {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.927940] env[65788]: DEBUG nova.compute.provider_tree [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.972312] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7466b110-4a16-4111-bd7e-8b408c3d97d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.980771] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Suspending the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1037.981017] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-a98058ec-5202-468d-a192-9c403fa7737d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.988883] env[65788]: DEBUG oslo_vmware.api [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1037.988883] env[65788]: value = "task-4662974" [ 1037.988883] env[65788]: _type = "Task" [ 1037.988883] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.997828] env[65788]: DEBUG oslo_vmware.api [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662974, 'name': SuspendVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.002041] env[65788]: WARNING neutronclient.v2_0.client [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1038.002555] env[65788]: WARNING openstack [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1038.002911] env[65788]: WARNING openstack [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1038.126208] env[65788]: DEBUG nova.network.neutron [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Updated VIF entry in instance network info cache for port b382ad0a-1833-409d-8d27-6726195bc924. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1038.126623] env[65788]: DEBUG nova.network.neutron [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Updating instance_info_cache with network_info: [{"id": "b382ad0a-1833-409d-8d27-6726195bc924", "address": "fa:16:3e:a2:c3:61", "network": {"id": "434cda43-035e-4e61-a5b8-0145c358571f", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-1972275156-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "828bafa0635a42c7be55b1fb367e1734", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb382ad0a-18", "ovs_interfaceid": "b382ad0a-1833-409d-8d27-6726195bc924", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1038.251829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.252301] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.252742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1038.253045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.253392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.255919] env[65788]: INFO nova.compute.manager [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Terminating instance [ 1038.432527] env[65788]: DEBUG nova.scheduler.client.report [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1038.499716] env[65788]: DEBUG oslo_vmware.api [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662974, 'name': SuspendVM_Task} progress is 58%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.629471] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9e716cd-17a5-422d-a474-b0b3b7359672 req-1cee3138-f399-430d-b597-f5fe5a560738 service nova] Releasing lock "refresh_cache-9d2f0d5e-793d-4790-ad31-be3d372be4fa" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.764850] env[65788]: DEBUG nova.compute.manager [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1038.764850] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1038.764850] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8642efbb-5e15-42a9-b01c-84fc5c7af9aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.775016] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1038.775016] env[65788]: value = "task-4662975" [ 1038.775016] env[65788]: _type = "Task" [ 1038.775016] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.792526] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.941152] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.942893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.503s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.944732] env[65788]: INFO nova.compute.claims [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.980070] env[65788]: INFO nova.scheduler.client.report [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Deleted allocations for instance 902d8a69-0398-4752-a609-0ac5b4b32e27 [ 1039.002892] env[65788]: DEBUG oslo_vmware.api [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662974, 'name': SuspendVM_Task, 'duration_secs': 0.78575} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.003633] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Suspended the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1039.003847] env[65788]: DEBUG nova.compute.manager [None req-dd2597b7-77a4-45fe-b298-0ddc08246c11 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1039.004770] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2d4c9d-9bb2-4260-a5b8-6f63241bd3a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.293181] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662975, 'name': PowerOffVM_Task, 'duration_secs': 0.309948} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.293549] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1039.293785] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1039.294052] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910355', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'name': 'volume-86b8defd-6720-439f-9a6c-b759609a2da3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '7f637326-9388-4d73-a1c7-3e4611ac46c5', 'attached_at': '2025-11-21T13:53:16.000000', 'detached_at': '', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'serial': '86b8defd-6720-439f-9a6c-b759609a2da3'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1039.294902] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc790621-4058-465c-b605-fcc1bf2d8884 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.319247] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdafdb8-c27d-4e29-8c8f-100f78195057 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.329703] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1afe2301-d5f2-4562-b7cc-c4c2cf0ae3ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.352773] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad84e72d-b7b9-48f9-810b-ef1368e9c177 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.369573] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The volume has not been displaced from its original location: [datastore2] volume-86b8defd-6720-439f-9a6c-b759609a2da3/volume-86b8defd-6720-439f-9a6c-b759609a2da3.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1039.374873] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1039.375244] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a157127-2c8d-4fa4-873c-ebaa31f399d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.395325] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1039.395325] env[65788]: value = "task-4662976" [ 1039.395325] env[65788]: _type = "Task" [ 1039.395325] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.404673] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.488781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dc8255cb-ed0b-44be-85e3-15084a9e8fc7 tempest-ServerDiskConfigTestJSON-664757069 tempest-ServerDiskConfigTestJSON-664757069-project-member] Lock "902d8a69-0398-4752-a609-0ac5b4b32e27" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.332s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1039.909978] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662976, 'name': ReconfigVM_Task, 'duration_secs': 0.253843} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.910351] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1039.916391] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07600c05-f0cf-4fd5-afea-9f8bb68cdf40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.933319] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1039.933319] env[65788]: value = "task-4662978" [ 1039.933319] env[65788]: _type = "Task" [ 1039.933319] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.944514] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662978, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.250209] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6299ff1-b8ce-4a3e-9159-ce050f03362c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.262137] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73405ad5-7ff9-4b71-8971-df2b005b25f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.297543] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334301f1-1e23-42af-b58b-316cbc7e6f26 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.306657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2c909f-5517-4678-ae77-fad721321259 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.324038] env[65788]: DEBUG nova.compute.provider_tree [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.445335] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662978, 'name': ReconfigVM_Task, 'duration_secs': 0.171128} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.445647] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910355', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'name': 'volume-86b8defd-6720-439f-9a6c-b759609a2da3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '7f637326-9388-4d73-a1c7-3e4611ac46c5', 'attached_at': '2025-11-21T13:53:16.000000', 'detached_at': '', 'volume_id': '86b8defd-6720-439f-9a6c-b759609a2da3', 'serial': '86b8defd-6720-439f-9a6c-b759609a2da3'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1040.446036] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1040.447411] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba49887-0e08-4e08-8bb4-e407b1b2b61b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.455934] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1040.456273] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02e7bf7d-3d19-4999-b00c-03a12282d707 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.524037] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1040.524258] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1040.524320] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleting the datastore file [datastore2] 7f637326-9388-4d73-a1c7-3e4611ac46c5 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.524574] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff1e6c6a-d67a-4759-a72e-51dd9e7f5a78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.532760] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1040.532760] env[65788]: value = "task-4662980" [ 1040.532760] env[65788]: _type = "Task" [ 1040.532760] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.542779] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.696206] env[65788]: INFO nova.compute.manager [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Resuming [ 1040.696269] env[65788]: DEBUG nova.objects.instance [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'flavor' on Instance uuid a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.827302] env[65788]: DEBUG nova.scheduler.client.report [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1041.045907] env[65788]: DEBUG oslo_vmware.api [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4662980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203401} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.046276] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1041.046386] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1041.046533] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1041.046710] env[65788]: INFO nova.compute.manager [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Took 2.28 seconds to destroy the instance on the hypervisor. [ 1041.046971] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1041.047198] env[65788]: DEBUG nova.compute.manager [-] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1041.047302] env[65788]: DEBUG nova.network.neutron [-] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1041.047579] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.048125] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.048382] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1041.332437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.333065] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1041.337463] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.354s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.337777] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.337966] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1041.338340] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.939s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.338564] env[65788]: DEBUG nova.objects.instance [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lazy-loading 'resources' on Instance uuid 9438ab56-1b4c-4778-a608-de319ab0ee43 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.345020] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d2819e-e526-41fa-8fd2-d2fc19858ec0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.356216] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705c567a-8c75-4581-bd5b-359c2f7dbe16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.375535] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea72657-6cd1-47b6-8d25-456143c1a611 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.384171] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502486c0-db84-462b-95db-3f9f8a27eea0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.425261] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179071MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1041.425261] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.442197] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.842188] env[65788]: DEBUG nova.compute.utils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1041.843719] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1041.843873] env[65788]: DEBUG nova.network.neutron [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1041.844806] env[65788]: WARNING neutronclient.v2_0.client [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.845912] env[65788]: WARNING neutronclient.v2_0.client [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1041.846854] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1041.847295] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.159687] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcc44b7-57b2-4f67-bfbf-a65b9ae3ffdb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.179582] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16eb8db-41a7-4f8c-a538-17674f0ddc5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.216683] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.216862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquired lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.217043] env[65788]: DEBUG nova.network.neutron [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1042.222691] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd95f35b-4c15-4c17-8ca1-0a0850338ed9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.235623] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98925991-d3a3-4aa1-a8d7-301d7dd43d63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.252817] env[65788]: DEBUG nova.compute.provider_tree [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.360948] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1042.589552] env[65788]: DEBUG nova.network.neutron [-] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1042.607996] env[65788]: DEBUG nova.policy [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32b4adcc8fa42fba0dbffd4582fcf69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36859c1eb994614b2a77400f811cf50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1042.722948] env[65788]: WARNING neutronclient.v2_0.client [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1042.723807] env[65788]: WARNING openstack [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1042.724474] env[65788]: WARNING openstack [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1042.756616] env[65788]: DEBUG nova.scheduler.client.report [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1043.084850] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1043.088907] env[65788]: DEBUG nova.network.neutron [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Successfully created port: c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1043.092652] env[65788]: INFO nova.compute.manager [-] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Took 2.04 seconds to deallocate network for instance. [ 1043.262783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.924s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1043.266136] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.067s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1043.266393] env[65788]: DEBUG nova.objects.instance [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lazy-loading 'resources' on Instance uuid 6e78042d-ba10-4629-816f-3a13b2e22d4e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.294361] env[65788]: INFO nova.scheduler.client.report [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Deleted allocations for instance 9438ab56-1b4c-4778-a608-de319ab0ee43 [ 1043.377016] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1043.408020] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1043.408020] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1043.408020] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1043.408524] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1043.410060] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1043.410060] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1043.410060] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.410060] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1043.410060] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1043.410541] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1043.410809] env[65788]: DEBUG nova.virt.hardware [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1043.411668] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040b5643-ae3f-41c2-9d46-6aa410be4e06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.423805] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995f2261-70ed-49d7-8769-19cada085d79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.653477] env[65788]: INFO nova.compute.manager [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Took 0.56 seconds to detach 1 volumes for instance. [ 1043.685472] env[65788]: WARNING openstack [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.685472] env[65788]: WARNING openstack [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1043.805146] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7fd01214-5567-48c5-bd50-8b0a01e0db50 tempest-ServerRescueNegativeTestJSON-1644419569 tempest-ServerRescueNegativeTestJSON-1644419569-project-member] Lock "9438ab56-1b4c-4778-a608-de319ab0ee43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.887s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1043.954054] env[65788]: DEBUG nova.compute.manager [req-c41e30f2-9f79-48e8-a31d-95e123e8126f req-a80cd102-3437-4170-907f-eec13aa85855 service nova] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Received event network-vif-deleted-e98f1b15-0a05-4be4-b950-dd7e6749eb8f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1043.958012] env[65788]: WARNING neutronclient.v2_0.client [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1043.960824] env[65788]: WARNING openstack [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1043.961499] env[65788]: WARNING openstack [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1044.057400] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705e5809-0af6-4717-aecf-1b5e5474e010 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.067533] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1553db-78a1-489e-849d-dc3eee2a0840 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.111261] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e95eef-5ab6-4fbd-a584-6b0f7ed8719b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.124217] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8035849f-23b2-4f4e-bb15-59fe46e268a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.141814] env[65788]: DEBUG nova.compute.provider_tree [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.164014] env[65788]: DEBUG nova.network.neutron [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [{"id": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "address": "fa:16:3e:0e:d1:43", "network": {"id": "54e56198-fde7-4b1b-9fc5-3f9463c18479", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1341023201-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7aa18dcc41403389518d700c2c8b8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cd6eb89-f7", "ovs_interfaceid": "2cd6eb89-f768-4ee6-93a2-386b83c42638", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1044.165853] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.560193] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.560573] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.461014] env[65788]: DEBUG nova.network.neutron [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Successfully updated port: c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1045.462708] env[65788]: DEBUG nova.scheduler.client.report [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1045.465920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Releasing lock "refresh_cache-a35ced42-4317-49b4-b4cc-4ed7e2c85c64" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.466338] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1045.470038] env[65788]: DEBUG nova.compute.manager [req-f3537ad7-7b2c-450b-8dd8-0171aad533f9 req-1c710d6a-ecf6-4a07-a5d5-be48377dc290 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Received event network-vif-plugged-c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1045.470275] env[65788]: DEBUG oslo_concurrency.lockutils [req-f3537ad7-7b2c-450b-8dd8-0171aad533f9 req-1c710d6a-ecf6-4a07-a5d5-be48377dc290 service nova] Acquiring lock "64acef8e-640d-4195-9519-ce31d984ddde-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1045.470497] env[65788]: DEBUG oslo_concurrency.lockutils [req-f3537ad7-7b2c-450b-8dd8-0171aad533f9 req-1c710d6a-ecf6-4a07-a5d5-be48377dc290 service nova] Lock "64acef8e-640d-4195-9519-ce31d984ddde-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.470670] env[65788]: DEBUG oslo_concurrency.lockutils [req-f3537ad7-7b2c-450b-8dd8-0171aad533f9 req-1c710d6a-ecf6-4a07-a5d5-be48377dc290 service nova] Lock "64acef8e-640d-4195-9519-ce31d984ddde-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.470835] env[65788]: DEBUG nova.compute.manager [req-f3537ad7-7b2c-450b-8dd8-0171aad533f9 req-1c710d6a-ecf6-4a07-a5d5-be48377dc290 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] No waiting events found dispatching network-vif-plugged-c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1045.470994] env[65788]: WARNING nova.compute.manager [req-f3537ad7-7b2c-450b-8dd8-0171aad533f9 req-1c710d6a-ecf6-4a07-a5d5-be48377dc290 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Received unexpected event network-vif-plugged-c4e814dc-0974-4f84-9c52-5feb163ebd78 for instance with vm_state building and task_state spawning. [ 1045.472521] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b531fda8-55ec-4464-a4b6-e6325f029e24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.481238] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Resuming the VM {{(pid=65788) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1045.481795] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ed5310b-a543-4a08-a0f0-a916cec7f30b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.491454] env[65788]: DEBUG oslo_vmware.api [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1045.491454] env[65788]: value = "task-4662981" [ 1045.491454] env[65788]: _type = "Task" [ 1045.491454] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.500967] env[65788]: DEBUG oslo_vmware.api [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662981, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.975349] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-64acef8e-640d-4195-9519-ce31d984ddde" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.975349] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-64acef8e-640d-4195-9519-ce31d984ddde" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.975349] env[65788]: DEBUG nova.network.neutron [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1045.975349] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.709s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1045.982226] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.972s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1045.982599] env[65788]: DEBUG nova.objects.instance [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lazy-loading 'resources' on Instance uuid 51705add-4d88-40bb-b236-2486650c77f4 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.009219] env[65788]: DEBUG oslo_vmware.api [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662981, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.010683] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.025755] env[65788]: INFO nova.scheduler.client.report [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted allocations for instance 6e78042d-ba10-4629-816f-3a13b2e22d4e [ 1046.483035] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.483523] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.509046] env[65788]: DEBUG oslo_vmware.api [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662981, 'name': PowerOnVM_Task, 'duration_secs': 0.537998} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.509349] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Resumed the VM {{(pid=65788) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1046.509525] env[65788]: DEBUG nova.compute.manager [None req-52b885d5-0332-4b6a-8e9e-acc10e51b5f6 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1046.510350] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee6816a-b2bc-468e-bcac-3a1014758071 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.538482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-02a43d71-8342-451f-9c26-61fc9ecb2fe0 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "6e78042d-ba10-4629-816f-3a13b2e22d4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.872s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.707091] env[65788]: DEBUG nova.network.neutron [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1046.745488] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558a7f50-47b3-432c-a3a0-ee40fce39234 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.750514] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.750901] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1046.762196] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184a81de-814b-4b1e-b5d4-44f4af56d75f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.797848] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83526d49-f870-42df-aebe-0d678593763e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.807571] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348b4ed8-faa6-42e6-b0c9-ef2fe0f81f33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.822958] env[65788]: DEBUG nova.compute.provider_tree [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.988684] env[65788]: WARNING neutronclient.v2_0.client [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1046.989392] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1046.989745] env[65788]: WARNING openstack [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1047.205718] env[65788]: DEBUG nova.network.neutron [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Updating instance_info_cache with network_info: [{"id": "c4e814dc-0974-4f84-9c52-5feb163ebd78", "address": "fa:16:3e:98:8a:d5", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e814dc-09", "ovs_interfaceid": "c4e814dc-0974-4f84-9c52-5feb163ebd78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1047.328094] env[65788]: DEBUG nova.scheduler.client.report [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1047.533602] env[65788]: DEBUG nova.compute.manager [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Received event network-changed-c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1047.534092] env[65788]: DEBUG nova.compute.manager [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Refreshing instance network info cache due to event network-changed-c4e814dc-0974-4f84-9c52-5feb163ebd78. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1047.534428] env[65788]: DEBUG oslo_concurrency.lockutils [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Acquiring lock "refresh_cache-64acef8e-640d-4195-9519-ce31d984ddde" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.710847] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-64acef8e-640d-4195-9519-ce31d984ddde" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.711268] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Instance network_info: |[{"id": "c4e814dc-0974-4f84-9c52-5feb163ebd78", "address": "fa:16:3e:98:8a:d5", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e814dc-09", "ovs_interfaceid": "c4e814dc-0974-4f84-9c52-5feb163ebd78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1047.712076] env[65788]: DEBUG oslo_concurrency.lockutils [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Acquired lock "refresh_cache-64acef8e-640d-4195-9519-ce31d984ddde" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.712076] env[65788]: DEBUG nova.network.neutron [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Refreshing network info cache for port c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1047.713528] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:8a:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4e814dc-0974-4f84-9c52-5feb163ebd78', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.721149] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1047.721758] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.721998] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f16444f-9aef-4bb3-b027-1280fab29324 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.744571] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.744571] env[65788]: value = "task-4662982" [ 1047.744571] env[65788]: _type = "Task" [ 1047.744571] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.754971] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662982, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.838554] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.840995] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.495s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.844406] env[65788]: INFO nova.compute.claims [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.858773] env[65788]: INFO nova.scheduler.client.report [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted allocations for instance 51705add-4d88-40bb-b236-2486650c77f4 [ 1048.077729] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.077729] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.077729] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.136924] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.136924] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.136924] env[65788]: INFO nova.compute.manager [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Terminating instance [ 1048.136924] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.136924] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.225150] env[65788]: WARNING neutronclient.v2_0.client [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.228326] env[65788]: WARNING openstack [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.229709] env[65788]: WARNING openstack [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.259638] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662982, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.367083] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ec99b11b-94b8-4be2-99f2-f7de50745cee tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "51705add-4d88-40bb-b236-2486650c77f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.543s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.550789] env[65788]: WARNING openstack [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.551373] env[65788]: WARNING openstack [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.590395] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1048.593521] env[65788]: DEBUG nova.compute.manager [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1048.593726] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.594955] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619793c3-023f-4996-9a4f-5f73fe135497 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.608032] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.608032] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c20da406-0a54-4ba2-b93b-601b5ef3b9d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.614866] env[65788]: DEBUG oslo_vmware.api [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1048.614866] env[65788]: value = "task-4662983" [ 1048.614866] env[65788]: _type = "Task" [ 1048.614866] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.618751] env[65788]: WARNING neutronclient.v2_0.client [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.619456] env[65788]: WARNING openstack [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1048.619817] env[65788]: WARNING openstack [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1048.631656] env[65788]: DEBUG oslo_vmware.api [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662983, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.740021] env[65788]: DEBUG nova.network.neutron [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Updated VIF entry in instance network info cache for port c4e814dc-0974-4f84-9c52-5feb163ebd78. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1048.740021] env[65788]: DEBUG nova.network.neutron [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Updating instance_info_cache with network_info: [{"id": "c4e814dc-0974-4f84-9c52-5feb163ebd78", "address": "fa:16:3e:98:8a:d5", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e814dc-09", "ovs_interfaceid": "c4e814dc-0974-4f84-9c52-5feb163ebd78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1048.759909] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662982, 'name': CreateVM_Task, 'duration_secs': 0.727651} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.761093] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.761093] env[65788]: WARNING neutronclient.v2_0.client [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1048.761093] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.761447] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.761620] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1048.761911] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffe53d26-cf94-4aaf-9f3e-003be27663a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.767996] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1048.767996] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52269b86-74a0-8e9c-8495-642ec2d28e6e" [ 1048.767996] env[65788]: _type = "Task" [ 1048.767996] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.777207] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52269b86-74a0-8e9c-8495-642ec2d28e6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.125526] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.129175] env[65788]: DEBUG oslo_vmware.api [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662983, 'name': PowerOffVM_Task, 'duration_secs': 0.369476} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.129358] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.129592] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.129879] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70c466b3-7941-4e31-9ce1-3bda0fc9aef2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.159040] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166b502c-54af-421b-80be-243ec22e0e93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.169729] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f8a5e4-00d1-4438-9c8b-6034aa7e652e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.207037] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3da0339-a8c5-4da9-8ebf-140c17b07d20 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.213063] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.213306] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.214891] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.215189] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.215372] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Deleting the datastore file [datastore1] d8103991-fd1f-4a57-81ce-1a47dc4defe0 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.217207] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79e834e5-0081-4c5a-8f9f-2c04e0b1013c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.224314] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "8755c56a-0f5e-4e95-a8b8-c643849fa798" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.224541] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.231570] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdde06b-6801-41a4-89cd-95ead7910299 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.239191] env[65788]: DEBUG oslo_vmware.api [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1049.239191] env[65788]: value = "task-4662985" [ 1049.239191] env[65788]: _type = "Task" [ 1049.239191] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.252166] env[65788]: DEBUG oslo_concurrency.lockutils [req-573144f6-a501-4828-bb13-30c8bf3d9162 req-0bd1c40b-6d29-4ba8-8425-4d78672894a9 service nova] Releasing lock "refresh_cache-64acef8e-640d-4195-9519-ce31d984ddde" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.253788] env[65788]: DEBUG nova.compute.provider_tree [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.261079] env[65788]: DEBUG oslo_vmware.api [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662985, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.265339] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "398791b2-9dc0-478a-8ed8-bdfbace0404e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1049.265560] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1049.282496] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52269b86-74a0-8e9c-8495-642ec2d28e6e, 'name': SearchDatastore_Task, 'duration_secs': 0.023757} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.282771] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.283896] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1049.283896] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.283896] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1049.283896] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1049.283896] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d45e1810-4cfb-46c2-a2cb-9cba152dbed7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.293368] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1049.293607] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1049.294345] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94e9bd7d-9feb-494b-b8ee-7874f7146f38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.301117] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1049.301117] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c2b0f9-a03f-6d7d-577e-1b33b7753f70" [ 1049.301117] env[65788]: _type = "Task" [ 1049.301117] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.309273] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c2b0f9-a03f-6d7d-577e-1b33b7753f70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.720071] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1049.744661] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1049.757790] env[65788]: DEBUG nova.scheduler.client.report [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1049.768023] env[65788]: DEBUG oslo_vmware.api [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4662985, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169872} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.768023] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.768023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.768023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.768023] env[65788]: INFO nova.compute.manager [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1049.768023] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1049.768023] env[65788]: DEBUG nova.compute.manager [-] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1049.768023] env[65788]: DEBUG nova.network.neutron [-] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1049.768023] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1049.768570] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1049.768746] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1049.776650] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1049.819966] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c2b0f9-a03f-6d7d-577e-1b33b7753f70, 'name': SearchDatastore_Task, 'duration_secs': 0.009186} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.822967] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abfd372a-dce6-4f3f-916d-43646478fb72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.842608] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1049.842608] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5260cc66-1afa-6487-f32b-073a985539bc" [ 1049.842608] env[65788]: _type = "Task" [ 1049.842608] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.856326] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5260cc66-1afa-6487-f32b-073a985539bc, 'name': SearchDatastore_Task, 'duration_secs': 0.010984} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.856773] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.857172] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 64acef8e-640d-4195-9519-ce31d984ddde/64acef8e-640d-4195-9519-ce31d984ddde.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.858520] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b1c5b36-3fb1-4f18-9a53-05d284f0670c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.868533] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1049.868533] env[65788]: value = "task-4662986" [ 1049.868533] env[65788]: _type = "Task" [ 1049.868533] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.881707] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.885808] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.163412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "2d74b823-6e28-444d-a80b-c91c2d595460" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.163767] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "2d74b823-6e28-444d-a80b-c91c2d595460" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.212538] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "a022cb81-5ff8-41aa-9307-c99a58416e94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.212690] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.256341] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.268591] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1050.269879] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1050.274468] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.274771] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.696s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1050.277747] env[65788]: INFO nova.compute.claims [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.303138] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1050.384573] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662986, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.483834] env[65788]: DEBUG nova.compute.manager [req-e6dfad28-b050-41ff-a198-205e8787de32 req-93229851-3d54-4167-a98c-35d3f2c53a53 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Received event network-vif-deleted-4ac3e7fa-b26e-4239-be70-ae1397fbba52 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1050.484058] env[65788]: INFO nova.compute.manager [req-e6dfad28-b050-41ff-a198-205e8787de32 req-93229851-3d54-4167-a98c-35d3f2c53a53 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Neutron deleted interface 4ac3e7fa-b26e-4239-be70-ae1397fbba52; detaching it from the instance and deleting it from the info cache [ 1050.484287] env[65788]: DEBUG nova.network.neutron [req-e6dfad28-b050-41ff-a198-205e8787de32 req-93229851-3d54-4167-a98c-35d3f2c53a53 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1050.666526] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1050.783841] env[65788]: DEBUG nova.compute.utils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1050.785499] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1050.785569] env[65788]: DEBUG nova.network.neutron [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1050.785865] env[65788]: WARNING neutronclient.v2_0.client [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.786228] env[65788]: WARNING neutronclient.v2_0.client [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1050.786819] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1050.787214] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1050.836430] env[65788]: DEBUG nova.policy [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4182220252a4ba5a384f67935c9a30a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3509ac857c94665a5aefc9e58c6f116', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1050.881752] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662986, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.942900] env[65788]: DEBUG nova.network.neutron [-] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1050.988102] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45736b9d-29ea-4f35-b076-b72cf2697ff2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.998502] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4071fc-2fd6-419e-8069-a3dfc45f3092 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.034477] env[65788]: DEBUG nova.compute.manager [req-e6dfad28-b050-41ff-a198-205e8787de32 req-93229851-3d54-4167-a98c-35d3f2c53a53 service nova] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Detach interface failed, port_id=4ac3e7fa-b26e-4239-be70-ae1397fbba52, reason: Instance d8103991-fd1f-4a57-81ce-1a47dc4defe0 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1051.174239] env[65788]: DEBUG nova.network.neutron [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Successfully created port: 92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1051.194889] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.298389] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1051.392155] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662986, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.447923] env[65788]: INFO nova.compute.manager [-] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Took 1.68 seconds to deallocate network for instance. [ 1051.617235] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b48d0b-e8dd-4160-afc7-3c5fbda2151a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.626146] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a845875-458a-4864-83b1-1ab581dbe478 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.660066] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58ed8db-5203-4c3f-902b-1e315d0bcac2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.668669] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a952c5f6-33dd-427b-9ec4-0bc3cb6bdcaf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.684213] env[65788]: DEBUG nova.compute.provider_tree [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.882627] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662986, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.553023} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.883123] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 64acef8e-640d-4195-9519-ce31d984ddde/64acef8e-640d-4195-9519-ce31d984ddde.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1051.883412] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1051.883694] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fcd361f7-d4cb-4a7f-b2d9-264c4ee5043d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.893472] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1051.893472] env[65788]: value = "task-4662987" [ 1051.893472] env[65788]: _type = "Task" [ 1051.893472] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.903202] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662987, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.958210] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.187313] env[65788]: DEBUG nova.scheduler.client.report [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1052.310455] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1052.344745] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1052.344872] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1052.345122] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1052.345223] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1052.345504] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1052.345504] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1052.345960] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.345960] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1052.349165] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1052.349165] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1052.349165] env[65788]: DEBUG nova.virt.hardware [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1052.349165] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55567b03-618e-4098-a438-53871d74065e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.357621] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9016318c-dae9-468f-b093-859f3ff8177d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.404260] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662987, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069046} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.404883] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1052.405534] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091bab62-710c-4a88-924a-4c71f57562ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.429233] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 64acef8e-640d-4195-9519-ce31d984ddde/64acef8e-640d-4195-9519-ce31d984ddde.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1052.429597] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa0631bc-5ae0-4da1-8660-3371a76b75d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.450665] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1052.450665] env[65788]: value = "task-4662988" [ 1052.450665] env[65788]: _type = "Task" [ 1052.450665] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.460046] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662988, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.626676] env[65788]: DEBUG nova.compute.manager [req-d3f29d8a-e5b5-4d49-b69c-863be2fe1d6b req-ad7f993c-c98d-4241-927d-7384ead83155 service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Received event network-vif-plugged-92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1052.626910] env[65788]: DEBUG oslo_concurrency.lockutils [req-d3f29d8a-e5b5-4d49-b69c-863be2fe1d6b req-ad7f993c-c98d-4241-927d-7384ead83155 service nova] Acquiring lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.627606] env[65788]: DEBUG oslo_concurrency.lockutils [req-d3f29d8a-e5b5-4d49-b69c-863be2fe1d6b req-ad7f993c-c98d-4241-927d-7384ead83155 service nova] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.627606] env[65788]: DEBUG oslo_concurrency.lockutils [req-d3f29d8a-e5b5-4d49-b69c-863be2fe1d6b req-ad7f993c-c98d-4241-927d-7384ead83155 service nova] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.627606] env[65788]: DEBUG nova.compute.manager [req-d3f29d8a-e5b5-4d49-b69c-863be2fe1d6b req-ad7f993c-c98d-4241-927d-7384ead83155 service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] No waiting events found dispatching network-vif-plugged-92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1052.627724] env[65788]: WARNING nova.compute.manager [req-d3f29d8a-e5b5-4d49-b69c-863be2fe1d6b req-ad7f993c-c98d-4241-927d-7384ead83155 service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Received unexpected event network-vif-plugged-92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 for instance with vm_state building and task_state spawning. [ 1052.693319] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.693893] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1052.697862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.313s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.699783] env[65788]: INFO nova.compute.claims [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.726550] env[65788]: DEBUG nova.network.neutron [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Successfully updated port: 92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1052.960963] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662988, 'name': ReconfigVM_Task, 'duration_secs': 0.306509} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.961331] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 64acef8e-640d-4195-9519-ce31d984ddde/64acef8e-640d-4195-9519-ce31d984ddde.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.961955] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddec1c19-10a5-42f8-bc89-23d2043518a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.969747] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1052.969747] env[65788]: value = "task-4662989" [ 1052.969747] env[65788]: _type = "Task" [ 1052.969747] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.979430] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662989, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.199524] env[65788]: DEBUG nova.compute.utils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1053.201113] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1053.202045] env[65788]: DEBUG nova.network.neutron [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1053.202045] env[65788]: WARNING neutronclient.v2_0.client [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.202045] env[65788]: WARNING neutronclient.v2_0.client [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.202629] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.202963] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.229598] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "refresh_cache-31a6981e-89cd-4b83-85cc-36dd163e1f8f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.229831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquired lock "refresh_cache-31a6981e-89cd-4b83-85cc-36dd163e1f8f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.230207] env[65788]: DEBUG nova.network.neutron [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1053.269602] env[65788]: DEBUG nova.policy [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c4a00121e214dc9a7b0866c8deee18e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ffbe44d63c1d432e97849f15615329e1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1053.482354] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662989, 'name': Rename_Task, 'duration_secs': 0.148239} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.482639] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1053.482893] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c707eed-2e5d-49d9-b2aa-21698b024864 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.490795] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1053.490795] env[65788]: value = "task-4662990" [ 1053.490795] env[65788]: _type = "Task" [ 1053.490795] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.505151] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662990, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.622440] env[65788]: DEBUG nova.network.neutron [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Successfully created port: ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1053.711252] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1053.736149] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.736709] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.783650] env[65788]: DEBUG nova.network.neutron [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1053.866440] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.866624] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1053.969041] env[65788]: WARNING neutronclient.v2_0.client [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1053.969749] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1053.970118] env[65788]: WARNING openstack [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1054.005086] env[65788]: DEBUG oslo_vmware.api [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662990, 'name': PowerOnVM_Task, 'duration_secs': 0.494477} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.005382] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1054.005583] env[65788]: INFO nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Took 10.63 seconds to spawn the instance on the hypervisor. [ 1054.005764] env[65788]: DEBUG nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1054.006578] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac835f8-1ece-4cb4-b7c8-816042e71c72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.066028] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3dbc7a-f0e5-4823-a604-90dbfabfef2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.073693] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7e556d-9080-43c4-b379-0bbd6c7e01de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.105936] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e19dfb-fd83-4706-9088-34beceacfb14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.116495] env[65788]: DEBUG nova.network.neutron [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Updating instance_info_cache with network_info: [{"id": "92c5d7f8-f9bf-43a9-b827-f02a33eba7d0", "address": "fa:16:3e:92:87:06", "network": {"id": "1b445a19-714a-4f18-94ca-506d14c5c09b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-270757226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3509ac857c94665a5aefc9e58c6f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c5d7f8-f9", "ovs_interfaceid": "92c5d7f8-f9bf-43a9-b827-f02a33eba7d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1054.118795] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b445f4a4-f5f7-4a3d-8ea0-53c57267a98b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.136158] env[65788]: DEBUG nova.compute.provider_tree [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.221736] env[65788]: INFO nova.virt.block_device [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Booting with volume 84d81198-187d-46ec-870e-36818081a75e at /dev/sda [ 1054.260535] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1159054-30cd-496c-bf3a-0a539f05ee49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.270606] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91013f22-7c3b-4c62-bbb3-348a9a6c0a5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.302607] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97096c00-78a3-4e5a-953c-29a377c32756 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.312938] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a916ac93-dfc7-44e0-977e-28f856463fa8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.343672] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8f1be3-a502-4582-b314-f14538f32546 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.351075] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a3efed-3c08-4a23-bc78-5d8cef3d5e5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.365947] env[65788]: DEBUG nova.virt.block_device [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating existing volume attachment record: ead2cda7-b237-446a-b183-2bc5eff516ed {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1054.480283] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.480616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.480861] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.481113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.481429] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.483968] env[65788]: INFO nova.compute.manager [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Terminating instance [ 1054.528316] env[65788]: INFO nova.compute.manager [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Took 35.13 seconds to build instance. [ 1054.626286] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Releasing lock "refresh_cache-31a6981e-89cd-4b83-85cc-36dd163e1f8f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1054.626703] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Instance network_info: |[{"id": "92c5d7f8-f9bf-43a9-b827-f02a33eba7d0", "address": "fa:16:3e:92:87:06", "network": {"id": "1b445a19-714a-4f18-94ca-506d14c5c09b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-270757226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3509ac857c94665a5aefc9e58c6f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c5d7f8-f9", "ovs_interfaceid": "92c5d7f8-f9bf-43a9-b827-f02a33eba7d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1054.627234] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:87:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92c5d7f8-f9bf-43a9-b827-f02a33eba7d0', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1054.636857] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Creating folder: Project (e3509ac857c94665a5aefc9e58c6f116). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1054.637226] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e6aeb0e1-8bb3-4310-9c18-bfa4a06fe2eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.642734] env[65788]: DEBUG nova.scheduler.client.report [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1054.657018] env[65788]: DEBUG nova.compute.manager [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Received event network-changed-92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1054.657256] env[65788]: DEBUG nova.compute.manager [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Refreshing instance network info cache due to event network-changed-92c5d7f8-f9bf-43a9-b827-f02a33eba7d0. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1054.657480] env[65788]: DEBUG oslo_concurrency.lockutils [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Acquiring lock "refresh_cache-31a6981e-89cd-4b83-85cc-36dd163e1f8f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.657622] env[65788]: DEBUG oslo_concurrency.lockutils [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Acquired lock "refresh_cache-31a6981e-89cd-4b83-85cc-36dd163e1f8f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1054.657810] env[65788]: DEBUG nova.network.neutron [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Refreshing network info cache for port 92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1054.661217] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Created folder: Project (e3509ac857c94665a5aefc9e58c6f116) in parent group-v910111. [ 1054.661321] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Creating folder: Instances. Parent ref: group-v910391. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1054.662436] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4fd19a2-8515-4291-a948-7f5fc06f2caa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.674990] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Created folder: Instances in parent group-v910391. [ 1054.675296] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1054.676435] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1054.680200] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7583a6df-43eb-468e-ad3d-29dbd8aebddd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.703599] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1054.703599] env[65788]: value = "task-4662993" [ 1054.703599] env[65788]: _type = "Task" [ 1054.703599] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.713606] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662993, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.988880] env[65788]: DEBUG nova.compute.manager [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1054.989101] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.990014] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd8611c-85da-4e4d-a2d9-53243fe505db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.998588] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.998877] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be31e577-b19d-422a-9825-53af511c1d33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.005978] env[65788]: DEBUG oslo_vmware.api [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1055.005978] env[65788]: value = "task-4662994" [ 1055.005978] env[65788]: _type = "Task" [ 1055.005978] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.016612] env[65788]: DEBUG oslo_vmware.api [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.030508] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1e56d52b-d381-4efa-b926-9a59a079800c tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.643s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.084480] env[65788]: DEBUG nova.compute.manager [req-c5d18e8b-d08b-4957-917d-fc88330d8f8d req-17d764aa-c5f3-42a2-ba8e-8e94624a8a06 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Received event network-vif-plugged-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1055.084480] env[65788]: DEBUG oslo_concurrency.lockutils [req-c5d18e8b-d08b-4957-917d-fc88330d8f8d req-17d764aa-c5f3-42a2-ba8e-8e94624a8a06 service nova] Acquiring lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.084480] env[65788]: DEBUG oslo_concurrency.lockutils [req-c5d18e8b-d08b-4957-917d-fc88330d8f8d req-17d764aa-c5f3-42a2-ba8e-8e94624a8a06 service nova] Lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.084480] env[65788]: DEBUG oslo_concurrency.lockutils [req-c5d18e8b-d08b-4957-917d-fc88330d8f8d req-17d764aa-c5f3-42a2-ba8e-8e94624a8a06 service nova] Lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.085394] env[65788]: DEBUG nova.compute.manager [req-c5d18e8b-d08b-4957-917d-fc88330d8f8d req-17d764aa-c5f3-42a2-ba8e-8e94624a8a06 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] No waiting events found dispatching network-vif-plugged-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1055.085394] env[65788]: WARNING nova.compute.manager [req-c5d18e8b-d08b-4957-917d-fc88330d8f8d req-17d764aa-c5f3-42a2-ba8e-8e94624a8a06 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Received unexpected event network-vif-plugged-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 for instance with vm_state building and task_state block_device_mapping. [ 1055.148598] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.148946] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1055.151823] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.727s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.163384] env[65788]: WARNING neutronclient.v2_0.client [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.163548] env[65788]: WARNING openstack [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.163999] env[65788]: WARNING openstack [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.192330] env[65788]: DEBUG nova.network.neutron [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Successfully updated port: ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1055.196172] env[65788]: DEBUG oslo_concurrency.lockutils [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "64acef8e-640d-4195-9519-ce31d984ddde" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.196172] env[65788]: DEBUG oslo_concurrency.lockutils [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.196172] env[65788]: DEBUG nova.compute.manager [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1055.198303] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd576ca1-2ffc-45b6-b6c1-2df54a3b4252 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.214767] env[65788]: DEBUG nova.compute.manager [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1055.215392] env[65788]: DEBUG nova.objects.instance [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'flavor' on Instance uuid 64acef8e-640d-4195-9519-ce31d984ddde {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.224049] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4662993, 'name': CreateVM_Task, 'duration_secs': 0.324388} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.224365] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1055.224819] env[65788]: WARNING neutronclient.v2_0.client [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.225229] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.225782] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.226160] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1055.226714] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c13e3d2e-46c4-4c6e-b149-2a6aba4e8711 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.232948] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1055.232948] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52120465-e40c-1b15-4570-c3fdc921d03a" [ 1055.232948] env[65788]: _type = "Task" [ 1055.232948] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.243074] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52120465-e40c-1b15-4570-c3fdc921d03a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.284948] env[65788]: WARNING openstack [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.285379] env[65788]: WARNING openstack [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.347931] env[65788]: WARNING neutronclient.v2_0.client [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.348666] env[65788]: WARNING openstack [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.349038] env[65788]: WARNING openstack [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.436571] env[65788]: DEBUG nova.network.neutron [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Updated VIF entry in instance network info cache for port 92c5d7f8-f9bf-43a9-b827-f02a33eba7d0. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1055.436944] env[65788]: DEBUG nova.network.neutron [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Updating instance_info_cache with network_info: [{"id": "92c5d7f8-f9bf-43a9-b827-f02a33eba7d0", "address": "fa:16:3e:92:87:06", "network": {"id": "1b445a19-714a-4f18-94ca-506d14c5c09b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-270757226-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3509ac857c94665a5aefc9e58c6f116", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92c5d7f8-f9", "ovs_interfaceid": "92c5d7f8-f9bf-43a9-b827-f02a33eba7d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1055.516875] env[65788]: DEBUG oslo_vmware.api [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662994, 'name': PowerOffVM_Task, 'duration_secs': 0.242373} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.517124] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.517291] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1055.517559] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d84c254-72cd-4006-b862-2c63d28fd6ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.533622] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1055.592218] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1055.592449] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1055.592630] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleting the datastore file [datastore2] a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1055.592902] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-981f0a38-e2d6-47d2-bb2e-3c98fd5c4f0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.600264] env[65788]: DEBUG oslo_vmware.api [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for the task: (returnval){ [ 1055.600264] env[65788]: value = "task-4662996" [ 1055.600264] env[65788]: _type = "Task" [ 1055.600264] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.610383] env[65788]: DEBUG oslo_vmware.api [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.656860] env[65788]: DEBUG nova.compute.utils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1055.664680] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1055.664902] env[65788]: DEBUG nova.network.neutron [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1055.665254] env[65788]: WARNING neutronclient.v2_0.client [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.665876] env[65788]: WARNING neutronclient.v2_0.client [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1055.666298] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1055.666795] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1055.674861] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1055.697525] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.697732] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.697994] env[65788]: DEBUG nova.network.neutron [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1055.730813] env[65788]: DEBUG nova.policy [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca13693271604f5790bdc8e19aa208a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7aa1dec53aa4436daa0f3223b92f9b79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1055.744727] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52120465-e40c-1b15-4570-c3fdc921d03a, 'name': SearchDatastore_Task, 'duration_secs': 0.010995} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.746048] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1055.746048] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1055.746048] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.746189] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1055.746418] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1055.746706] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b1f614a-8e36-49b4-a3ae-dd50579170ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.757779] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1055.758045] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1055.758880] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e144e41e-4fae-4db5-9e99-01cf4211ac14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.767161] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1055.767161] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a43e65-9acc-6881-c708-947073d837f3" [ 1055.767161] env[65788]: _type = "Task" [ 1055.767161] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.782809] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a43e65-9acc-6881-c708-947073d837f3, 'name': SearchDatastore_Task, 'duration_secs': 0.010836} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.784198] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2e6f7ff-d8d2-4b22-9b69-6f24fc0d7079 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.792295] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1055.792295] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5238a80a-006d-3cff-fb87-5c29f7dba687" [ 1055.792295] env[65788]: _type = "Task" [ 1055.792295] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.806684] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5238a80a-006d-3cff-fb87-5c29f7dba687, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.940266] env[65788]: DEBUG oslo_concurrency.lockutils [req-6a6c09ac-b747-4de5-8e25-c7bad4c3ba39 req-69022b3c-c75d-4bf4-bfaf-049b7644a5fc service nova] Releasing lock "refresh_cache-31a6981e-89cd-4b83-85cc-36dd163e1f8f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1056.062036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.069792] env[65788]: DEBUG nova.network.neutron [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Successfully created port: d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1056.111711] env[65788]: DEBUG oslo_vmware.api [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Task: {'id': task-4662996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15813} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.112018] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.112223] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1056.112416] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1056.112631] env[65788]: INFO nova.compute.manager [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1056.112848] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1056.113864] env[65788]: DEBUG nova.compute.manager [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1056.113864] env[65788]: DEBUG nova.network.neutron [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1056.113864] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.113986] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.114528] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.169040] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.200808] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.200808] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d8103991-fd1f-4a57-81ce-1a47dc4defe0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1056.200808] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance ec7f7a46-8dd2-4b75-866f-20e73907f1cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.200808] env[65788]: WARNING nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 7f637326-9388-4d73-a1c7-3e4611ac46c5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1056.200808] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1a701758-a056-4948-9069-2a7168f2cc8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.201059] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance a35ced42-4317-49b4-b4cc-4ed7e2c85c64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.201059] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9d2f0d5e-793d-4790-ad31-be3d372be4fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.201143] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 64acef8e-640d-4195-9519-ce31d984ddde actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.203024] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.203731] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.222431] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.222431] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dae3919-3401-407d-841f-09a02cfb9587 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.231042] env[65788]: DEBUG oslo_vmware.api [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1056.231042] env[65788]: value = "task-4662997" [ 1056.231042] env[65788]: _type = "Task" [ 1056.231042] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.244443] env[65788]: DEBUG oslo_vmware.api [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.310043] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5238a80a-006d-3cff-fb87-5c29f7dba687, 'name': SearchDatastore_Task, 'duration_secs': 0.011593} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.310790] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1056.311078] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 31a6981e-89cd-4b83-85cc-36dd163e1f8f/31a6981e-89cd-4b83-85cc-36dd163e1f8f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1056.311788] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6863cbc7-f82a-476e-81f6-ac2b2aac97d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.324578] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1056.324578] env[65788]: value = "task-4662998" [ 1056.324578] env[65788]: _type = "Task" [ 1056.324578] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.336056] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4662998, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.437538] env[65788]: DEBUG nova.network.neutron [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1056.458436] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1056.459137] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1056.459411] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1056.459584] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1056.460179] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1056.460430] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1056.460682] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1056.460995] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.461275] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1056.461617] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1056.461881] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1056.462180] env[65788]: DEBUG nova.virt.hardware [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1056.463285] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c837d92b-8340-4f0c-93b3-5f0ed234aec5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.474317] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e3a7a8-3d32-4e37-a81d-3eaccfd17fad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.507930] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.508374] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.620230] env[65788]: WARNING neutronclient.v2_0.client [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1056.620921] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1056.621325] env[65788]: WARNING openstack [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1056.688652] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1056.711942] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4dcf8335-701d-4386-9de4-f14f5d375d1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1056.712154] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 31a6981e-89cd-4b83-85cc-36dd163e1f8f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.712278] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 18be7322-3359-49c9-a181-f9228d5c16d7 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.712375] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance bc39a402-1f5a-4a67-b09f-a18159cf4abf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1056.721822] env[65788]: DEBUG nova.network.neutron [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1056.726253] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1056.726591] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1056.726795] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1056.727050] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1056.727268] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1056.727479] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1056.727688] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.727850] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1056.728205] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1056.728414] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1056.728636] env[65788]: DEBUG nova.virt.hardware [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1056.729779] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40199924-c9f1-4065-8ab3-e04ceb053f2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.754413] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f161db3-6448-42ab-aa61-111c14cbeedb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.761730] env[65788]: DEBUG oslo_vmware.api [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4662997, 'name': PowerOffVM_Task, 'duration_secs': 0.243058} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.762136] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.762447] env[65788]: DEBUG nova.compute.manager [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1056.764331] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6917fae-a860-4ded-9d6a-8cc044c653ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.834407] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4662998, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509179} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.834709] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 31a6981e-89cd-4b83-85cc-36dd163e1f8f/31a6981e-89cd-4b83-85cc-36dd163e1f8f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.834955] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.835239] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ff2dff2-a438-422d-a229-d744a8ea7132 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.841814] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1056.841814] env[65788]: value = "task-4662999" [ 1056.841814] env[65788]: _type = "Task" [ 1056.841814] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.852033] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4662999, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.907134] env[65788]: DEBUG nova.network.neutron [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.180850] env[65788]: DEBUG nova.compute.manager [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Received event network-changed-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1057.180914] env[65788]: DEBUG nova.compute.manager [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Refreshing instance network info cache due to event network-changed-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1057.181114] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Acquiring lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.215189] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1057.236528] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.236927] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Instance network_info: |[{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1057.237292] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Acquired lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1057.237469] env[65788]: DEBUG nova.network.neutron [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Refreshing network info cache for port ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1057.239034] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:f5:65', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '099fe970-c61f-4480-bed4-ae4f485fd82a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.247474] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1057.248808] env[65788]: WARNING neutronclient.v2_0.client [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.249441] env[65788]: WARNING openstack [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.249789] env[65788]: WARNING openstack [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.256946] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1057.257811] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d82c9ca5-2cc3-44ed-967f-ec711c7d6ae7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.280397] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.280397] env[65788]: value = "task-4663000" [ 1057.280397] env[65788]: _type = "Task" [ 1057.280397] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.289597] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663000, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.296700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-83e980f4-3eac-4644-95ce-418c7218e667 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.101s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.353527] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4662999, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06307} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.353936] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.355090] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-888c6eae-784b-436c-a215-eab56c5b0189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.379706] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 31a6981e-89cd-4b83-85cc-36dd163e1f8f/31a6981e-89cd-4b83-85cc-36dd163e1f8f.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.380076] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37819c05-9494-4e3b-ba97-c5ec00b0f33d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.402907] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1057.402907] env[65788]: value = "task-4663001" [ 1057.402907] env[65788]: _type = "Task" [ 1057.402907] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.412558] env[65788]: INFO nova.compute.manager [-] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Took 1.30 seconds to deallocate network for instance. [ 1057.413389] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663001, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.444019] env[65788]: WARNING openstack [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.444019] env[65788]: WARNING openstack [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.586322] env[65788]: WARNING neutronclient.v2_0.client [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.587145] env[65788]: WARNING openstack [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1057.587549] env[65788]: WARNING openstack [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1057.698433] env[65788]: DEBUG nova.network.neutron [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Successfully updated port: d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1057.718611] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1057.744130] env[65788]: DEBUG nova.network.neutron [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updated VIF entry in instance network info cache for port ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1057.744130] env[65788]: DEBUG nova.network.neutron [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1057.748268] env[65788]: DEBUG nova.compute.manager [req-4fcff8aa-bad2-484b-88b0-6a862d5e50cc req-facfc45d-0e21-4054-b867-a7a7006f66f0 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Received event network-vif-plugged-d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1057.748898] env[65788]: DEBUG oslo_concurrency.lockutils [req-4fcff8aa-bad2-484b-88b0-6a862d5e50cc req-facfc45d-0e21-4054-b867-a7a7006f66f0 service nova] Acquiring lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.748898] env[65788]: DEBUG oslo_concurrency.lockutils [req-4fcff8aa-bad2-484b-88b0-6a862d5e50cc req-facfc45d-0e21-4054-b867-a7a7006f66f0 service nova] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.749045] env[65788]: DEBUG oslo_concurrency.lockutils [req-4fcff8aa-bad2-484b-88b0-6a862d5e50cc req-facfc45d-0e21-4054-b867-a7a7006f66f0 service nova] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.749272] env[65788]: DEBUG nova.compute.manager [req-4fcff8aa-bad2-484b-88b0-6a862d5e50cc req-facfc45d-0e21-4054-b867-a7a7006f66f0 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] No waiting events found dispatching network-vif-plugged-d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1057.750410] env[65788]: WARNING nova.compute.manager [req-4fcff8aa-bad2-484b-88b0-6a862d5e50cc req-facfc45d-0e21-4054-b867-a7a7006f66f0 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Received unexpected event network-vif-plugged-d6d70bc9-e4ba-4775-9452-626cf247086d for instance with vm_state building and task_state spawning. [ 1057.794394] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663000, 'name': CreateVM_Task, 'duration_secs': 0.401888} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.794551] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.795213] env[65788]: WARNING neutronclient.v2_0.client [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1057.795598] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'ead2cda7-b237-446a-b183-2bc5eff516ed', 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910380', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'name': 'volume-84d81198-187d-46ec-870e-36818081a75e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '18be7322-3359-49c9-a181-f9228d5c16d7', 'attached_at': '', 'detached_at': '', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'serial': '84d81198-187d-46ec-870e-36818081a75e'}, 'disk_bus': None, 'device_type': None, 'delete_on_termination': True, 'mount_device': '/dev/sda', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65788) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1057.795828] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Root volume attach. Driver type: vmdk {{(pid=65788) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1057.796989] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3dfeaa-dab3-45f1-ae7f-dd0871ef68ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.807040] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed517966-1a74-4ba7-a685-f58590de1ac3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.814596] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2dcd23b-0000-4ac6-a1e6-613c9dbedaff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.824128] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-96220885-c3f2-4eb5-ba97-1d2a6f416df2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.832512] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1057.832512] env[65788]: value = "task-4663002" [ 1057.832512] env[65788]: _type = "Task" [ 1057.832512] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.848712] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663002, 'name': RelocateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.913749] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663001, 'name': ReconfigVM_Task, 'duration_secs': 0.300823} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.914129] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 31a6981e-89cd-4b83-85cc-36dd163e1f8f/31a6981e-89cd-4b83-85cc-36dd163e1f8f.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.914933] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce33da0d-1394-4ab7-8d8a-bfbc10f6afb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.923615] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.924041] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1057.924041] env[65788]: value = "task-4663003" [ 1057.924041] env[65788]: _type = "Task" [ 1057.924041] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.934965] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663003, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.936488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "64acef8e-640d-4195-9519-ce31d984ddde" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.936705] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.936901] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "64acef8e-640d-4195-9519-ce31d984ddde-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.937092] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.937257] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.939197] env[65788]: INFO nova.compute.manager [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Terminating instance [ 1058.203070] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.204233] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.204233] env[65788]: DEBUG nova.network.neutron [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1058.222087] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance c5dfbb9b-430d-4e93-b24d-e918d90e123e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1058.247228] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] Releasing lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.247629] env[65788]: DEBUG nova.compute.manager [req-4c58cfef-90dc-4544-a6e6-44b5b4a5a25c req-3285681b-1275-41b2-9557-9b298c52636b service nova] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Received event network-vif-deleted-2cd6eb89-f768-4ee6-93a2-386b83c42638 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1058.343968] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663002, 'name': RelocateVM_Task, 'duration_secs': 0.438729} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.344308] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1058.344519] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910380', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'name': 'volume-84d81198-187d-46ec-870e-36818081a75e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '18be7322-3359-49c9-a181-f9228d5c16d7', 'attached_at': '', 'detached_at': '', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'serial': '84d81198-187d-46ec-870e-36818081a75e'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1058.345362] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f584bc-5f7b-49d6-a8f2-136325decf42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.364149] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba68bf71-eaa6-4755-927b-9b5c421bf42d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.388490] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-84d81198-187d-46ec-870e-36818081a75e/volume-84d81198-187d-46ec-870e-36818081a75e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1058.388689] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b09117c8-f4cb-40e4-8f94-282ef7abf8d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.410034] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1058.410034] env[65788]: value = "task-4663004" [ 1058.410034] env[65788]: _type = "Task" [ 1058.410034] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.419902] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.435291] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663003, 'name': Rename_Task, 'duration_secs': 0.185837} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.435582] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1058.435982] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-252c6fdb-7a2e-44be-a7bf-91063befaa25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.446186] env[65788]: DEBUG nova.compute.manager [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1058.446469] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1058.446840] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1058.446840] env[65788]: value = "task-4663005" [ 1058.446840] env[65788]: _type = "Task" [ 1058.446840] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.448098] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b3d7fb-c381-401a-b655-1f1ed3ef1138 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.463089] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663005, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.466229] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1058.466647] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72e5914d-9efb-44c8-89e9-9121000c32e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.706542] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.706932] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.726568] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 8755c56a-0f5e-4e95-a8b8-c643849fa798 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1058.731144] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1058.731405] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1058.731609] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleting the datastore file [datastore2] 64acef8e-640d-4195-9519-ce31d984ddde {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.731918] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8511257e-b1c1-48de-8de8-fb9ec366c4e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.740946] env[65788]: DEBUG oslo_vmware.api [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1058.740946] env[65788]: value = "task-4663007" [ 1058.740946] env[65788]: _type = "Task" [ 1058.740946] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.753536] env[65788]: DEBUG oslo_vmware.api [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.754804] env[65788]: DEBUG nova.network.neutron [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1058.779583] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.779979] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.856238] env[65788]: WARNING neutronclient.v2_0.client [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1058.857417] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1058.858008] env[65788]: WARNING openstack [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1058.924546] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663004, 'name': ReconfigVM_Task, 'duration_secs': 0.336792} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.924921] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-84d81198-187d-46ec-870e-36818081a75e/volume-84d81198-187d-46ec-870e-36818081a75e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1058.930632] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f022c33-9735-40a8-a98d-bfd456213992 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.949728] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1058.949728] env[65788]: value = "task-4663008" [ 1058.949728] env[65788]: _type = "Task" [ 1058.949728] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.965017] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663005, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.968860] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.973674] env[65788]: DEBUG nova.network.neutron [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updating instance_info_cache with network_info: [{"id": "d6d70bc9-e4ba-4775-9452-626cf247086d", "address": "fa:16:3e:46:8a:7d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d70bc9-e4", "ovs_interfaceid": "d6d70bc9-e4ba-4775-9452-626cf247086d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1059.229977] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 398791b2-9dc0-478a-8ed8-bdfbace0404e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1059.254580] env[65788]: DEBUG oslo_vmware.api [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195166} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.254993] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.255067] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1059.255199] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1059.255366] env[65788]: INFO nova.compute.manager [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Took 0.81 seconds to destroy the instance on the hypervisor. [ 1059.255620] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1059.255818] env[65788]: DEBUG nova.compute.manager [-] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1059.255910] env[65788]: DEBUG nova.network.neutron [-] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1059.256166] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.256688] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1059.256945] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1059.294899] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1059.466239] env[65788]: DEBUG oslo_vmware.api [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663005, 'name': PowerOnVM_Task, 'duration_secs': 0.60067} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.466239] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663008, 'name': ReconfigVM_Task, 'duration_secs': 0.155784} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.466463] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.466660] env[65788]: INFO nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Took 7.16 seconds to spawn the instance on the hypervisor. [ 1059.466830] env[65788]: DEBUG nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1059.467215] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910380', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'name': 'volume-84d81198-187d-46ec-870e-36818081a75e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '18be7322-3359-49c9-a181-f9228d5c16d7', 'attached_at': '', 'detached_at': '', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'serial': '84d81198-187d-46ec-870e-36818081a75e'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1059.468239] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0059d7c3-bd11-4dc6-ad48-50f50f926781 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.470679] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-648ec09a-abd7-40fd-8e9b-654761549977 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.476573] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.477023] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Instance network_info: |[{"id": "d6d70bc9-e4ba-4775-9452-626cf247086d", "address": "fa:16:3e:46:8a:7d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d70bc9-e4", "ovs_interfaceid": "d6d70bc9-e4ba-4775-9452-626cf247086d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1059.481284] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:8a:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6d70bc9-e4ba-4775-9452-626cf247086d', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.489115] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1059.489426] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1059.489426] env[65788]: value = "task-4663009" [ 1059.489426] env[65788]: _type = "Task" [ 1059.489426] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.489961] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.490255] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa2b58a5-fcb3-44c8-91ab-c4b64424d476 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.516211] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663009, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.517676] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.517676] env[65788]: value = "task-4663010" [ 1059.517676] env[65788]: _type = "Task" [ 1059.517676] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.527463] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663010, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.734469] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 2d74b823-6e28-444d-a80b-c91c2d595460 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1059.784231] env[65788]: DEBUG nova.compute.manager [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Received event network-changed-d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1059.784532] env[65788]: DEBUG nova.compute.manager [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Refreshing instance network info cache due to event network-changed-d6d70bc9-e4ba-4775-9452-626cf247086d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1059.784797] env[65788]: DEBUG oslo_concurrency.lockutils [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Acquiring lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.784979] env[65788]: DEBUG oslo_concurrency.lockutils [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Acquired lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1059.785161] env[65788]: DEBUG nova.network.neutron [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Refreshing network info cache for port d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1060.000816] env[65788]: INFO nova.compute.manager [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Took 31.68 seconds to build instance. [ 1060.008629] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663009, 'name': Rename_Task, 'duration_secs': 0.170674} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.008907] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.009354] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-792bb261-90fb-4ee8-9f88-813d128c929a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.019351] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1060.019351] env[65788]: value = "task-4663011" [ 1060.019351] env[65788]: _type = "Task" [ 1060.019351] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.037561] env[65788]: DEBUG nova.network.neutron [-] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1060.038959] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.039208] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663010, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.238017] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance a022cb81-5ff8-41aa-9307-c99a58416e94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1060.238334] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1060.238509] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=100GB used_disk=8GB total_vcpus=48 used_vcpus=9 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '9', 'num_vm_active': '6', 'num_task_None': '5', 'num_os_type_None': '9', 'num_proj_ffbe44d63c1d432e97849f15615329e1': '3', 'io_workload': '3', 'num_task_deleting': '1', 'num_proj_5e7aa18dcc41403389518d700c2c8b8a': '1', 'num_proj_7aa1dec53aa4436daa0f3223b92f9b79': '2', 'num_proj_828bafa0635a42c7be55b1fb367e1734': '1', 'num_task_powering-off': '1', 'num_proj_e36859c1eb994614b2a77400f811cf50': '1', 'num_vm_building': '3', 'num_task_spawning': '1', 'num_proj_e3509ac857c94665a5aefc9e58c6f116': '1', 'num_task_block_device_mapping': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1060.288041] env[65788]: WARNING neutronclient.v2_0.client [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.288878] env[65788]: WARNING openstack [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.289345] env[65788]: WARNING openstack [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.413475] env[65788]: WARNING openstack [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.414062] env[65788]: WARNING openstack [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.503489] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de562ae7-c03c-48ec-8f40-18ab3f156c21 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.189s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.509246] env[65788]: WARNING neutronclient.v2_0.client [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.510013] env[65788]: WARNING openstack [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1060.510382] env[65788]: WARNING openstack [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1060.533187] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663010, 'name': CreateVM_Task, 'duration_secs': 0.679177} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.538728] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.539197] env[65788]: DEBUG oslo_vmware.api [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663011, 'name': PowerOnVM_Task, 'duration_secs': 0.493356} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.539907] env[65788]: WARNING neutronclient.v2_0.client [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1060.540348] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.540575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.540950] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1060.541338] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.541641] env[65788]: INFO nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Took 4.08 seconds to spawn the instance on the hypervisor. [ 1060.541846] env[65788]: DEBUG nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1060.542380] env[65788]: INFO nova.compute.manager [-] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Took 1.29 seconds to deallocate network for instance. [ 1060.542653] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c83f5708-30ab-42c7-8191-0e89d1c51a42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.544972] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9571e471-48f5-45fe-888d-c4bf27097698 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.562279] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1060.562279] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f5d8b9-d538-32e6-633c-e83be932e1ca" [ 1060.562279] env[65788]: _type = "Task" [ 1060.562279] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.577749] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f5d8b9-d538-32e6-633c-e83be932e1ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.579560] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c091de57-3ec9-44f9-b99e-34f61c44504d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.590244] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a30bd7-fc45-4ee3-8c24-f052c471c8ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.629053] env[65788]: DEBUG nova.network.neutron [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updated VIF entry in instance network info cache for port d6d70bc9-e4ba-4775-9452-626cf247086d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1060.629460] env[65788]: DEBUG nova.network.neutron [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updating instance_info_cache with network_info: [{"id": "d6d70bc9-e4ba-4775-9452-626cf247086d", "address": "fa:16:3e:46:8a:7d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d70bc9-e4", "ovs_interfaceid": "d6d70bc9-e4ba-4775-9452-626cf247086d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1060.633132] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2605ba9-27d2-46b0-9885-6a6536b0946f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.643863] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b321662a-3aa8-460a-95f8-75517c15f7e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.661062] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.064524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.074846] env[65788]: INFO nova.compute.manager [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Took 32.51 seconds to build instance. [ 1061.081139] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f5d8b9-d538-32e6-633c-e83be932e1ca, 'name': SearchDatastore_Task, 'duration_secs': 0.025906} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.082395] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1061.082395] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.082395] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.082555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.082717] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.083932] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e0e6a8d-a18c-4137-a171-a2630291217f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.093559] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.093827] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.094706] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff3d698d-0c77-4b54-921e-a5eda5a86e1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.101446] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1061.101446] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5247e943-225a-2c94-5a68-274c71969a38" [ 1061.101446] env[65788]: _type = "Task" [ 1061.101446] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.110605] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5247e943-225a-2c94-5a68-274c71969a38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.137094] env[65788]: DEBUG oslo_concurrency.lockutils [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] Releasing lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1061.137392] env[65788]: DEBUG nova.compute.manager [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Received event network-vif-deleted-c4e814dc-0974-4f84-9c52-5feb163ebd78 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1061.137586] env[65788]: INFO nova.compute.manager [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Neutron deleted interface c4e814dc-0974-4f84-9c52-5feb163ebd78; detaching it from the instance and deleting it from the info cache [ 1061.137867] env[65788]: DEBUG nova.network.neutron [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1061.164066] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1061.259361] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.259509] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.259811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1061.261018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.261018] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.262843] env[65788]: INFO nova.compute.manager [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Terminating instance [ 1061.576663] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84bde57a-6cea-4ae3-98f5-99247cd46f18 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.022s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.616036] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5247e943-225a-2c94-5a68-274c71969a38, 'name': SearchDatastore_Task, 'duration_secs': 0.012123} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.616036] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-518c2cac-1d61-4567-b3a3-66e20b6549bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.622719] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1061.622719] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bb06e6-245b-1e1b-f617-a494b4e2a1b1" [ 1061.622719] env[65788]: _type = "Task" [ 1061.622719] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.633814] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bb06e6-245b-1e1b-f617-a494b4e2a1b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.641298] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d7e40b5-b99e-413f-b068-36eb18f2b3c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.651312] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7969e32b-3c64-4218-a5fc-5298d9f21d5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.669815] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1061.670077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.518s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1061.670389] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.586s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1061.670595] env[65788]: DEBUG nova.objects.instance [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'pci_requests' on Instance uuid 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.685572] env[65788]: DEBUG nova.compute.manager [req-78326cc6-0678-4a3a-b94d-30a11cc32bba req-bf8c199f-099e-45a3-bc2b-c905686c2190 service nova] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Detach interface failed, port_id=c4e814dc-0974-4f84-9c52-5feb163ebd78, reason: Instance 64acef8e-640d-4195-9519-ce31d984ddde could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1061.767257] env[65788]: DEBUG nova.compute.manager [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1061.767412] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.768231] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d91a6292-a66d-45fb-bd65-80ccd96ddf0f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.777275] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.777688] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71217be0-3a84-4994-943f-f0e5e4d36005 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.786302] env[65788]: DEBUG oslo_vmware.api [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1061.786302] env[65788]: value = "task-4663012" [ 1061.786302] env[65788]: _type = "Task" [ 1061.786302] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.795011] env[65788]: DEBUG oslo_vmware.api [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.042373] env[65788]: DEBUG nova.compute.manager [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Received event network-changed-e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1062.042373] env[65788]: DEBUG nova.compute.manager [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Refreshing instance network info cache due to event network-changed-e936d0e7-de51-4ac2-9b4a-08f999624c5c. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1062.042373] env[65788]: DEBUG oslo_concurrency.lockutils [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Acquiring lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.042373] env[65788]: DEBUG oslo_concurrency.lockutils [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Acquired lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1062.042373] env[65788]: DEBUG nova.network.neutron [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Refreshing network info cache for port e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1062.135344] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52bb06e6-245b-1e1b-f617-a494b4e2a1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.012914} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.135533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.135800] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] bc39a402-1f5a-4a67-b09f-a18159cf4abf/bc39a402-1f5a-4a67-b09f-a18159cf4abf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.136094] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77c4cbfb-3aee-4a0b-ae6f-ca947468eef9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.145708] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1062.145708] env[65788]: value = "task-4663013" [ 1062.145708] env[65788]: _type = "Task" [ 1062.145708] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.160045] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.175340] env[65788]: DEBUG nova.objects.instance [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'numa_topology' on Instance uuid 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.296717] env[65788]: DEBUG oslo_vmware.api [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663012, 'name': PowerOffVM_Task, 'duration_secs': 0.197732} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.297048] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1062.297227] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1062.298424] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e786a329-5788-4669-86ba-e2874dfcce19 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.422567] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1062.422835] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1062.423013] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Deleting the datastore file [datastore2] 31a6981e-89cd-4b83-85cc-36dd163e1f8f {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1062.423334] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d87c5f7-1203-48af-b08c-848cdb8cbfc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.433768] env[65788]: DEBUG oslo_vmware.api [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for the task: (returnval){ [ 1062.433768] env[65788]: value = "task-4663015" [ 1062.433768] env[65788]: _type = "Task" [ 1062.433768] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.449136] env[65788]: DEBUG oslo_vmware.api [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663015, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.549627] env[65788]: WARNING neutronclient.v2_0.client [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.550419] env[65788]: WARNING openstack [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.550798] env[65788]: WARNING openstack [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.656982] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663013, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.658454] env[65788]: WARNING openstack [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.658833] env[65788]: WARNING openstack [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.679941] env[65788]: INFO nova.compute.claims [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1062.724655] env[65788]: WARNING neutronclient.v2_0.client [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.726064] env[65788]: WARNING openstack [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.726064] env[65788]: WARNING openstack [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.841016] env[65788]: DEBUG nova.network.neutron [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updated VIF entry in instance network info cache for port e936d0e7-de51-4ac2-9b4a-08f999624c5c. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1062.842054] env[65788]: DEBUG nova.network.neutron [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updating instance_info_cache with network_info: [{"id": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "address": "fa:16:3e:ff:eb:0c", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape936d0e7-de", "ovs_interfaceid": "e936d0e7-de51-4ac2-9b4a-08f999624c5c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1062.945034] env[65788]: DEBUG oslo_vmware.api [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Task: {'id': task-4663015, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.382727} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.945238] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.945422] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.945599] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.945770] env[65788]: INFO nova.compute.manager [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1062.946035] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1062.946319] env[65788]: DEBUG nova.compute.manager [-] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1062.946425] env[65788]: DEBUG nova.network.neutron [-] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1062.946678] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1062.947217] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1062.947479] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1062.985949] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1063.160087] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534619} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.160087] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] bc39a402-1f5a-4a67-b09f-a18159cf4abf/bc39a402-1f5a-4a67-b09f-a18159cf4abf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.160087] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.160087] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db48a50a-a0c9-409e-a271-df1d59b3ef64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.168641] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1063.168641] env[65788]: value = "task-4663016" [ 1063.168641] env[65788]: _type = "Task" [ 1063.168641] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.180405] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.347184] env[65788]: DEBUG oslo_concurrency.lockutils [req-21ff59f0-4c35-4d61-9904-ddf2f1a85522 req-58feec2c-ee3a-4d3b-ada4-fb5f5ac1d582 service nova] Releasing lock "refresh_cache-efe1048b-50e9-4add-910a-607a95759c7a" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.532396] env[65788]: DEBUG nova.compute.manager [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1063.679627] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0793} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.680012] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.680694] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914fa848-e177-49ae-a6d8-b28b98925a29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.707019] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] bc39a402-1f5a-4a67-b09f-a18159cf4abf/bc39a402-1f5a-4a67-b09f-a18159cf4abf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.707758] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c89073e-962d-4a2b-abd0-87f7c4be048a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.735715] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1063.735715] env[65788]: value = "task-4663017" [ 1063.735715] env[65788]: _type = "Task" [ 1063.735715] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.748383] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.792719] env[65788]: DEBUG nova.network.neutron [-] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1063.996920] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b19b57-83e2-4df3-ad44-6a20d5d6df14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.005468] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40a7030-d1de-4fe8-b4a0-720c158762fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.041141] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8c46f8-6c39-4df4-b319-f89656e44782 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.049375] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68770a40-6d6a-4616-8b45-91f8c8755909 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.054580] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.065754] env[65788]: DEBUG nova.compute.provider_tree [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.074252] env[65788]: DEBUG nova.compute.manager [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Received event network-changed-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1064.074483] env[65788]: DEBUG nova.compute.manager [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Refreshing instance network info cache due to event network-changed-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1064.074752] env[65788]: DEBUG oslo_concurrency.lockutils [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Acquiring lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.074900] env[65788]: DEBUG oslo_concurrency.lockutils [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Acquired lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.075077] env[65788]: DEBUG nova.network.neutron [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Refreshing network info cache for port ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1064.247932] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663017, 'name': ReconfigVM_Task, 'duration_secs': 0.310603} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.247932] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Reconfigured VM instance instance-0000005f to attach disk [datastore2] bc39a402-1f5a-4a67-b09f-a18159cf4abf/bc39a402-1f5a-4a67-b09f-a18159cf4abf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.248333] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f8dc2ee-cfc8-49be-886f-a7051952c153 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.255852] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1064.255852] env[65788]: value = "task-4663018" [ 1064.255852] env[65788]: _type = "Task" [ 1064.255852] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.268735] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663018, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.295579] env[65788]: INFO nova.compute.manager [-] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Took 1.35 seconds to deallocate network for instance. [ 1064.569541] env[65788]: DEBUG nova.scheduler.client.report [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.577914] env[65788]: WARNING neutronclient.v2_0.client [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.578526] env[65788]: WARNING openstack [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.578869] env[65788]: WARNING openstack [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.709332] env[65788]: WARNING openstack [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.709332] env[65788]: WARNING openstack [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.766036] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663018, 'name': Rename_Task, 'duration_secs': 0.147597} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.766339] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.766606] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-577dbb63-7eb8-4322-bfff-cf6b0119b20b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.775125] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1064.775125] env[65788]: value = "task-4663019" [ 1064.775125] env[65788]: _type = "Task" [ 1064.775125] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.781098] env[65788]: WARNING neutronclient.v2_0.client [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1064.781781] env[65788]: WARNING openstack [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1064.782160] env[65788]: WARNING openstack [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1064.795974] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663019, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.803451] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1064.864418] env[65788]: DEBUG nova.network.neutron [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updated VIF entry in instance network info cache for port ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1064.864795] env[65788]: DEBUG nova.network.neutron [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1065.074677] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.404s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.075132] env[65788]: WARNING neutronclient.v2_0.client [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1065.077908] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.912s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.078155] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.080801] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.070s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.082471] env[65788]: INFO nova.compute.claims [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1065.109186] env[65788]: INFO nova.network.neutron [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating port eba2c4b7-13de-4fe3-95fb-86d9494cd97a with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1065.116910] env[65788]: INFO nova.scheduler.client.report [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted allocations for instance 7f637326-9388-4d73-a1c7-3e4611ac46c5 [ 1065.286031] env[65788]: DEBUG oslo_vmware.api [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663019, 'name': PowerOnVM_Task, 'duration_secs': 0.440072} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.286237] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.286447] env[65788]: INFO nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Took 8.60 seconds to spawn the instance on the hypervisor. [ 1065.286632] env[65788]: DEBUG nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1065.287469] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f4c4e7-09db-409a-9aa7-1d053138596e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.368166] env[65788]: DEBUG oslo_concurrency.lockutils [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] Releasing lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.368443] env[65788]: DEBUG nova.compute.manager [req-2ccdf9a6-b971-43fb-bb7d-a681a2571fe4 req-cff192b4-08a8-4364-950d-c4505cf0cc41 service nova] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Received event network-vif-deleted-92c5d7f8-f9bf-43a9-b827-f02a33eba7d0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1065.626949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8c5e77b6-bfa8-4a71-92e6-ac3f7bbf26ef tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "7f637326-9388-4d73-a1c7-3e4611ac46c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.375s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.806776] env[65788]: INFO nova.compute.manager [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Took 28.45 seconds to build instance. [ 1066.314187] env[65788]: DEBUG oslo_concurrency.lockutils [None req-642ffcbd-3e65-4d9b-9896-2dedfd81e07a tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.979s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.369071] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325cc0ed-ff5a-4610-8012-f44a53116143 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.378518] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-344c594f-80c0-439f-8594-86c5d62567cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.411560] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5fb138-a62b-478e-8a5e-911503037b07 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.420386] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7262f42-c746-4ed6-bfb2-38b4a3e53f0a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.435830] env[65788]: DEBUG nova.compute.provider_tree [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.642317] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.642555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.683017] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.683085] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.683268] env[65788]: DEBUG nova.network.neutron [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1066.803236] env[65788]: DEBUG nova.compute.manager [req-e2b9eea8-4c16-4614-8216-fb2c218ed8a9 req-fdb862aa-9643-4a89-b0e2-f1da5bd3963c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-vif-plugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1066.803390] env[65788]: DEBUG oslo_concurrency.lockutils [req-e2b9eea8-4c16-4614-8216-fb2c218ed8a9 req-fdb862aa-9643-4a89-b0e2-f1da5bd3963c service nova] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.803633] env[65788]: DEBUG oslo_concurrency.lockutils [req-e2b9eea8-4c16-4614-8216-fb2c218ed8a9 req-fdb862aa-9643-4a89-b0e2-f1da5bd3963c service nova] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.803808] env[65788]: DEBUG oslo_concurrency.lockutils [req-e2b9eea8-4c16-4614-8216-fb2c218ed8a9 req-fdb862aa-9643-4a89-b0e2-f1da5bd3963c service nova] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.803986] env[65788]: DEBUG nova.compute.manager [req-e2b9eea8-4c16-4614-8216-fb2c218ed8a9 req-fdb862aa-9643-4a89-b0e2-f1da5bd3963c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] No waiting events found dispatching network-vif-plugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1066.804888] env[65788]: WARNING nova.compute.manager [req-e2b9eea8-4c16-4614-8216-fb2c218ed8a9 req-fdb862aa-9643-4a89-b0e2-f1da5bd3963c service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received unexpected event network-vif-plugged-eba2c4b7-13de-4fe3-95fb-86d9494cd97a for instance with vm_state shelved_offloaded and task_state spawning. [ 1066.877143] env[65788]: DEBUG nova.compute.manager [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Received event network-changed-d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1066.877382] env[65788]: DEBUG nova.compute.manager [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Refreshing instance network info cache due to event network-changed-d6d70bc9-e4ba-4775-9452-626cf247086d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1066.877556] env[65788]: DEBUG oslo_concurrency.lockutils [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Acquiring lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.877697] env[65788]: DEBUG oslo_concurrency.lockutils [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Acquired lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.877855] env[65788]: DEBUG nova.network.neutron [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Refreshing network info cache for port d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1066.938768] env[65788]: DEBUG nova.scheduler.client.report [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1067.146059] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1067.186879] env[65788]: WARNING neutronclient.v2_0.client [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.187586] env[65788]: WARNING openstack [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.187935] env[65788]: WARNING openstack [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.327101] env[65788]: WARNING openstack [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.327583] env[65788]: WARNING openstack [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.381212] env[65788]: WARNING neutronclient.v2_0.client [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.382392] env[65788]: WARNING openstack [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.382836] env[65788]: WARNING openstack [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.396357] env[65788]: WARNING neutronclient.v2_0.client [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.397106] env[65788]: WARNING openstack [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.397472] env[65788]: WARNING openstack [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.443536] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.444131] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1067.447367] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.322s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.448838] env[65788]: INFO nova.compute.claims [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1067.506833] env[65788]: DEBUG nova.network.neutron [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1067.523155] env[65788]: WARNING openstack [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.523612] env[65788]: WARNING openstack [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.586896] env[65788]: WARNING neutronclient.v2_0.client [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.587608] env[65788]: WARNING openstack [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.587959] env[65788]: WARNING openstack [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.670496] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.673696] env[65788]: DEBUG nova.network.neutron [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updated VIF entry in instance network info cache for port d6d70bc9-e4ba-4775-9452-626cf247086d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1067.674167] env[65788]: DEBUG nova.network.neutron [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updating instance_info_cache with network_info: [{"id": "d6d70bc9-e4ba-4775-9452-626cf247086d", "address": "fa:16:3e:46:8a:7d", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6d70bc9-e4", "ovs_interfaceid": "d6d70bc9-e4ba-4775-9452-626cf247086d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1067.957783] env[65788]: DEBUG nova.compute.utils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1067.958997] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1067.960852] env[65788]: DEBUG nova.network.neutron [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1067.960852] env[65788]: WARNING neutronclient.v2_0.client [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.960852] env[65788]: WARNING neutronclient.v2_0.client [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1067.960852] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1067.960852] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1067.970548] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1068.009740] env[65788]: DEBUG nova.policy [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1068.010985] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.043550] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9f3a22ba8dfa3d23b5b6613c66afcd29',container_format='bare',created_at=2025-11-21T13:52:40Z,direct_url=,disk_format='vmdk',id=17403798-a0dc-480e-ac18-3ae8959fda51,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1098315683-shelved',owner='94977b75768444138fb914cabddc57c5',properties=ImageMetaProps,protected=,size=31671296,status='active',tags=,updated_at=2025-11-21T13:52:57Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1068.043871] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1068.044064] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1068.044261] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1068.044448] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1068.044561] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1068.044776] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.044997] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1068.045111] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1068.045280] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1068.045457] env[65788]: DEBUG nova.virt.hardware [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1068.046373] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e427c3e-9913-4722-8ec8-8c24530a6ba9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.055286] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43230fd-6306-4f4f-aa5e-bc1dabe08a5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.070329] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:23:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a0a76279-3c11-4bef-b124-2a2ee13fa377', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eba2c4b7-13de-4fe3-95fb-86d9494cd97a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1068.077996] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1068.078284] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1068.078517] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d2687c5-f81e-4070-a55f-2f8a5f0530a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.100124] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1068.100124] env[65788]: value = "task-4663020" [ 1068.100124] env[65788]: _type = "Task" [ 1068.100124] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.109597] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663020, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.177526] env[65788]: DEBUG oslo_concurrency.lockutils [req-ae38a51d-09ee-4c99-838a-31f201012d40 req-0ecb2a83-1aed-4f45-b5fc-741b10641dc6 service nova] Releasing lock "refresh_cache-bc39a402-1f5a-4a67-b09f-a18159cf4abf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.363998] env[65788]: DEBUG nova.network.neutron [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Successfully created port: 1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1068.609648] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663020, 'name': CreateVM_Task, 'duration_secs': 0.379083} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.609841] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1068.612713] env[65788]: WARNING neutronclient.v2_0.client [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1068.613110] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.613273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.613716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1068.614177] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8775f629-322a-4a75-a60f-e8e675ac3f03 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.619553] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1068.619553] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5259358b-5063-2abd-8fff-1b24df2eff5f" [ 1068.619553] env[65788]: _type = "Task" [ 1068.619553] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.630257] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5259358b-5063-2abd-8fff-1b24df2eff5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.732012] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ebb1b0-e562-4d65-82f2-fe2aaa1bca4a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.739900] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e5a148-7915-4bb0-826a-8943e3141526 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.771027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5740b565-f19f-4383-9e17-b73ad63fbecd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.779726] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc139ff-f3aa-41fe-ba61-ede869563fbd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.794859] env[65788]: DEBUG nova.compute.provider_tree [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.862834] env[65788]: DEBUG nova.compute.manager [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1068.863052] env[65788]: DEBUG nova.compute.manager [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing instance network info cache due to event network-changed-eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1068.863285] env[65788]: DEBUG oslo_concurrency.lockutils [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Acquiring lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.863439] env[65788]: DEBUG oslo_concurrency.lockutils [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Acquired lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.863613] env[65788]: DEBUG nova.network.neutron [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Refreshing network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1068.984057] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1069.010653] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1069.010905] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1069.011076] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1069.011270] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1069.011511] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1069.011701] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1069.011915] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.012090] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1069.012262] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1069.012428] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1069.012607] env[65788]: DEBUG nova.virt.hardware [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1069.013528] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdcefe6-dd97-431a-986d-ffddc38b0294 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.022163] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccd573e-d2d8-4997-9787-7bac0a968544 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.130346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1069.131090] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Processing image 17403798-a0dc-480e-ac18-3ae8959fda51 {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1069.131090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.131090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1069.131223] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1069.131581] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06225b6d-1bb7-4dc0-9711-17e9bd8caa94 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.142145] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1069.142325] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1069.143098] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cd8c5d6-dbc3-4663-9b75-c1503b5c076d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.148681] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1069.148681] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528bce45-89d3-bde9-4fc9-71104b1f54f0" [ 1069.148681] env[65788]: _type = "Task" [ 1069.148681] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.156942] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528bce45-89d3-bde9-4fc9-71104b1f54f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.298694] env[65788]: DEBUG nova.scheduler.client.report [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1069.367035] env[65788]: WARNING neutronclient.v2_0.client [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.367796] env[65788]: WARNING openstack [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.368220] env[65788]: WARNING openstack [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.498010] env[65788]: WARNING openstack [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.498523] env[65788]: WARNING openstack [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.562606] env[65788]: WARNING neutronclient.v2_0.client [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1069.563377] env[65788]: WARNING openstack [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1069.563808] env[65788]: WARNING openstack [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1069.659827] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1069.660179] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Fetch image to [datastore2] OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef/OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1069.660296] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Downloading stream optimized image 17403798-a0dc-480e-ac18-3ae8959fda51 to [datastore2] OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef/OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef.vmdk on the data store datastore2 as vApp {{(pid=65788) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1069.660469] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Downloading image file data 17403798-a0dc-480e-ac18-3ae8959fda51 to the ESX as VM named 'OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef' {{(pid=65788) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1069.734656] env[65788]: DEBUG nova.network.neutron [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updated VIF entry in instance network info cache for port eba2c4b7-13de-4fe3-95fb-86d9494cd97a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1069.735050] env[65788]: DEBUG nova.network.neutron [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [{"id": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "address": "fa:16:3e:91:23:a5", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeba2c4b7-13", "ovs_interfaceid": "eba2c4b7-13de-4fe3-95fb-86d9494cd97a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1069.753698] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1069.753698] env[65788]: value = "resgroup-9" [ 1069.753698] env[65788]: _type = "ResourcePool" [ 1069.753698] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1069.754011] env[65788]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-bf5e8657-4ab4-45a1-875f-6045754ba1e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.777755] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease: (returnval){ [ 1069.777755] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c14bc3-a548-d8fc-7728-bf460eeb3615" [ 1069.777755] env[65788]: _type = "HttpNfcLease" [ 1069.777755] env[65788]: } obtained for vApp import into resource pool (val){ [ 1069.777755] env[65788]: value = "resgroup-9" [ 1069.777755] env[65788]: _type = "ResourcePool" [ 1069.777755] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1069.778215] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the lease: (returnval){ [ 1069.778215] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c14bc3-a548-d8fc-7728-bf460eeb3615" [ 1069.778215] env[65788]: _type = "HttpNfcLease" [ 1069.778215] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1069.785678] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1069.785678] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c14bc3-a548-d8fc-7728-bf460eeb3615" [ 1069.785678] env[65788]: _type = "HttpNfcLease" [ 1069.785678] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1069.805137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.805731] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1069.809000] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.553s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.810874] env[65788]: INFO nova.compute.claims [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.820474] env[65788]: DEBUG nova.compute.manager [req-c583d5b7-34ed-4f0c-a87b-226a155efb65 req-4d0ee6e8-b97c-4d2f-9ee4-1cd378c744e9 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-vif-plugged-1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1069.820747] env[65788]: DEBUG oslo_concurrency.lockutils [req-c583d5b7-34ed-4f0c-a87b-226a155efb65 req-4d0ee6e8-b97c-4d2f-9ee4-1cd378c744e9 service nova] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.821025] env[65788]: DEBUG oslo_concurrency.lockutils [req-c583d5b7-34ed-4f0c-a87b-226a155efb65 req-4d0ee6e8-b97c-4d2f-9ee4-1cd378c744e9 service nova] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.821680] env[65788]: DEBUG oslo_concurrency.lockutils [req-c583d5b7-34ed-4f0c-a87b-226a155efb65 req-4d0ee6e8-b97c-4d2f-9ee4-1cd378c744e9 service nova] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.821680] env[65788]: DEBUG nova.compute.manager [req-c583d5b7-34ed-4f0c-a87b-226a155efb65 req-4d0ee6e8-b97c-4d2f-9ee4-1cd378c744e9 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] No waiting events found dispatching network-vif-plugged-1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1069.821680] env[65788]: WARNING nova.compute.manager [req-c583d5b7-34ed-4f0c-a87b-226a155efb65 req-4d0ee6e8-b97c-4d2f-9ee4-1cd378c744e9 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received unexpected event network-vif-plugged-1426fd1f-0507-498d-9a05-9bce1efd861a for instance with vm_state building and task_state spawning. [ 1069.925966] env[65788]: DEBUG nova.network.neutron [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Successfully updated port: 1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1070.238511] env[65788]: DEBUG oslo_concurrency.lockutils [req-8de14414-f60d-4ab3-89ba-0da01bdc1eb3 req-24def6fc-2a6b-404e-a364-2c66908f4a30 service nova] Releasing lock "refresh_cache-4dcf8335-701d-4386-9de4-f14f5d375d1f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.287415] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1070.287415] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c14bc3-a548-d8fc-7728-bf460eeb3615" [ 1070.287415] env[65788]: _type = "HttpNfcLease" [ 1070.287415] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1070.287633] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1070.287633] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c14bc3-a548-d8fc-7728-bf460eeb3615" [ 1070.287633] env[65788]: _type = "HttpNfcLease" [ 1070.287633] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1070.288206] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91996b28-dae4-44fc-aab2-3f6a3f7dc7c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.296019] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52715686-ca0f-ca1e-f1d3-80bd82456aa0/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1070.296155] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating HTTP connection to write to file with size = 31671296 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52715686-ca0f-ca1e-f1d3-80bd82456aa0/disk-0.vmdk. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1070.353463] env[65788]: DEBUG nova.compute.utils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1070.359538] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1070.359751] env[65788]: DEBUG nova.network.neutron [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1070.360172] env[65788]: WARNING neutronclient.v2_0.client [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.360416] env[65788]: WARNING neutronclient.v2_0.client [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1070.360995] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.361359] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.377225] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-70105f47-5c66-4eff-ba12-8d4278b9f114 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.420868] env[65788]: DEBUG nova.policy [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02eddcb388a431abdff72ad3995dda1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69c8fb9f477e45068cd7c1e4defb64a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1070.428861] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.429085] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1070.429300] env[65788]: DEBUG nova.network.neutron [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1070.767177] env[65788]: DEBUG nova.network.neutron [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Successfully created port: e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1070.856926] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1070.933120] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1070.933978] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1070.982610] env[65788]: DEBUG nova.network.neutron [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1071.005169] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.005169] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.119641] env[65788]: WARNING neutronclient.v2_0.client [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1071.120513] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1071.123637] env[65788]: WARNING openstack [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1071.239114] env[65788]: DEBUG nova.network.neutron [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1071.258335] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89449eb1-88d9-4dbf-b44f-706d79f42401 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.270367] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c54d19c-607b-4be9-b158-1feffe38629f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.308787] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80d65ae-9595-4cff-99bd-e4ab0541e640 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.321302] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d515208b-0097-40aa-9e1e-82e5ade9108c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.339505] env[65788]: DEBUG nova.compute.provider_tree [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.549411] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1071.549411] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52715686-ca0f-ca1e-f1d3-80bd82456aa0/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1071.550596] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9c4d04-b3b2-466d-a56a-54ae82b15213 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.559295] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52715686-ca0f-ca1e-f1d3-80bd82456aa0/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1071.559589] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52715686-ca0f-ca1e-f1d3-80bd82456aa0/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1071.559885] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-b76b867e-63f4-45a1-b5c6-2b06e068ef8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.741638] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1071.742133] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Instance network_info: |[{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1071.743311] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:4b:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1426fd1f-0507-498d-9a05-9bce1efd861a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.751764] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Creating folder: Project (676963c25a3741998b10790689242dbb). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1071.752453] env[65788]: DEBUG oslo_vmware.rw_handles [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52715686-ca0f-ca1e-f1d3-80bd82456aa0/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1071.752698] env[65788]: INFO nova.virt.vmwareapi.images [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Downloaded image file data 17403798-a0dc-480e-ac18-3ae8959fda51 [ 1071.752989] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aafd6bed-f9d7-4c4f-929a-ada80dffdd55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.756060] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb73266b-2ead-4a8d-b7b0-23baf54b4fdd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.776465] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38bd7d8d-d647-4024-bb47-8ed95e13af36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.778194] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Created folder: Project (676963c25a3741998b10790689242dbb) in parent group-v910111. [ 1071.778383] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Creating folder: Instances. Parent ref: group-v910398. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1071.778658] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b05d6a4-956a-4a64-a785-73b95c88bdb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.790874] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Created folder: Instances in parent group-v910398. [ 1071.791142] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1071.791417] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.791683] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3db9e833-9ddc-4553-b497-7cabe54f13e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.808212] env[65788]: INFO nova.virt.vmwareapi.images [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] The imported VM was unregistered [ 1071.810970] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1071.811239] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating directory with path [datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51 {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.812065] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3af30e72-957b-4cd4-b1ea-5dc179bc1ee8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.817191] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.817191] env[65788]: value = "task-4663025" [ 1071.817191] env[65788]: _type = "Task" [ 1071.817191] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.826749] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663025, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.828278] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created directory with path [datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51 {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.828608] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef/OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef.vmdk to [datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk. {{(pid=65788) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1071.828892] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-06ddeab5-4f6e-437b-a21e-858f6e3c5118 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.836311] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1071.836311] env[65788]: value = "task-4663026" [ 1071.836311] env[65788]: _type = "Task" [ 1071.836311] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.846027] env[65788]: DEBUG nova.scheduler.client.report [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1071.849634] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.852149] env[65788]: DEBUG nova.compute.manager [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-changed-1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1071.852439] env[65788]: DEBUG nova.compute.manager [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Refreshing instance network info cache due to event network-changed-1426fd1f-0507-498d-9a05-9bce1efd861a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1071.852607] env[65788]: DEBUG oslo_concurrency.lockutils [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Acquiring lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.852772] env[65788]: DEBUG oslo_concurrency.lockutils [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Acquired lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1071.853035] env[65788]: DEBUG nova.network.neutron [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Refreshing network info cache for port 1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1071.875522] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1071.906028] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1071.906028] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1071.906028] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1071.906028] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1071.906210] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1071.906344] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1071.906559] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1071.907279] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1071.907486] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1071.907660] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1071.907839] env[65788]: DEBUG nova.virt.hardware [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1071.908760] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0224a8-748c-493d-91ec-c0e85023ed16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.917724] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eead2dc-4518-4b97-9a02-4a7c0c2a1be7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.331731] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663025, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.351042] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.352075] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.352816] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1072.356558] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.082s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.358046] env[65788]: INFO nova.compute.claims [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1072.361248] env[65788]: WARNING neutronclient.v2_0.client [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.362131] env[65788]: WARNING openstack [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.363045] env[65788]: WARNING openstack [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.386107] env[65788]: DEBUG nova.network.neutron [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Successfully updated port: e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1072.498699] env[65788]: WARNING openstack [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.499130] env[65788]: WARNING openstack [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.567205] env[65788]: WARNING neutronclient.v2_0.client [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.567892] env[65788]: WARNING openstack [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.568261] env[65788]: WARNING openstack [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.655502] env[65788]: DEBUG nova.network.neutron [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updated VIF entry in instance network info cache for port 1426fd1f-0507-498d-9a05-9bce1efd861a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1072.656113] env[65788]: DEBUG nova.network.neutron [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1072.837670] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663025, 'name': CreateVM_Task, 'duration_secs': 0.70429} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.842599] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1072.843371] env[65788]: WARNING neutronclient.v2_0.client [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.843887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.844155] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1072.844600] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1072.845559] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1664ee0-f300-4da2-97eb-bdb7548baddd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.854656] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.856741] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1072.856741] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5297afde-7437-9c09-25fa-59a759df82c8" [ 1072.856741] env[65788]: _type = "Task" [ 1072.856741] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.869241] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5297afde-7437-9c09-25fa-59a759df82c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.876129] env[65788]: DEBUG nova.compute.utils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1072.878333] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1072.878600] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1072.879227] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.879619] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1072.880541] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1072.881061] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1072.895125] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1072.899727] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.899852] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1072.899987] env[65788]: DEBUG nova.network.neutron [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1072.934467] env[65788]: DEBUG nova.policy [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd335bd92f4f647c6b7ec8678b11b5613', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d8d51f6df554bcba9bd0e8dbbbcc3d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1073.159937] env[65788]: DEBUG oslo_concurrency.lockutils [req-dcaef193-ef24-465b-9358-67514ead546a req-82cb2f42-47ca-4d05-925d-ab43251e4fc0 service nova] Releasing lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1073.311086] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Successfully created port: 2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1073.349604] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.371943] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5297afde-7437-9c09-25fa-59a759df82c8, 'name': SearchDatastore_Task, 'duration_secs': 0.089098} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.372346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1073.372804] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.373616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.373616] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1073.373745] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.373955] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f0b585c-e221-4ec4-805e-31075711a53b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.396950] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.397342] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.399719] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b993081a-54ca-42d3-9196-96238d392c24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.408189] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.408627] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.423208] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1073.423208] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae421f-d86c-e958-720b-42b7736b970d" [ 1073.423208] env[65788]: _type = "Task" [ 1073.423208] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.433903] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae421f-d86c-e958-720b-42b7736b970d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.458081] env[65788]: DEBUG nova.network.neutron [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1073.506098] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.506484] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.589865] env[65788]: WARNING neutronclient.v2_0.client [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1073.590575] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1073.591228] env[65788]: WARNING openstack [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1073.727095] env[65788]: DEBUG nova.network.neutron [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1073.730463] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e6ede1-05de-47f5-abe7-fd25f449cb1e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.740440] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d63c4bc-13ea-45f0-bc66-b4d3ccb36a62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.775642] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abb62bb-54e4-48f1-8262-437b613b4112 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.785079] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0adb574-dd7f-4116-9975-9189f35cae89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.801355] env[65788]: DEBUG nova.compute.provider_tree [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.847964] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.881385] env[65788]: DEBUG nova.compute.manager [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Received event network-vif-plugged-e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1073.882105] env[65788]: DEBUG oslo_concurrency.lockutils [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.882105] env[65788]: DEBUG oslo_concurrency.lockutils [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1073.882225] env[65788]: DEBUG oslo_concurrency.lockutils [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1073.882403] env[65788]: DEBUG nova.compute.manager [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] No waiting events found dispatching network-vif-plugged-e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1073.882602] env[65788]: WARNING nova.compute.manager [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Received unexpected event network-vif-plugged-e7eece2d-f24b-4fcf-b1d8-a232af2c0021 for instance with vm_state building and task_state spawning. [ 1073.882842] env[65788]: DEBUG nova.compute.manager [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Received event network-changed-e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1073.883087] env[65788]: DEBUG nova.compute.manager [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Refreshing instance network info cache due to event network-changed-e7eece2d-f24b-4fcf-b1d8-a232af2c0021. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1073.883558] env[65788]: DEBUG oslo_concurrency.lockutils [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.919839] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1073.935237] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ae421f-d86c-e958-720b-42b7736b970d, 'name': SearchDatastore_Task, 'duration_secs': 0.088438} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.936226] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5903ec65-e38c-47e5-acf9-8d8599f36d8f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.944912] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1073.944912] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d62226-4111-fcd2-181c-57c4574c4798" [ 1073.944912] env[65788]: _type = "Task" [ 1073.944912] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.951570] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1073.951852] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1073.952014] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1073.952250] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1073.952513] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1073.952734] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1073.952961] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1073.953153] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1073.953430] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1073.953664] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1073.953858] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1073.955525] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f07715-4401-44c4-ae72-3844e7ca6783 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.964874] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d62226-4111-fcd2-181c-57c4574c4798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.968838] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a528c15d-138f-4a37-91ac-5d35d9191832 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.235411] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1074.235879] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Instance network_info: |[{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1074.236265] env[65788]: DEBUG oslo_concurrency.lockutils [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1074.236484] env[65788]: DEBUG nova.network.neutron [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Refreshing network info cache for port e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1074.237974] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:c5:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7eece2d-f24b-4fcf-b1d8-a232af2c0021', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1074.250647] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating folder: Project (69c8fb9f477e45068cd7c1e4defb64a0). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1074.252811] env[65788]: WARNING neutronclient.v2_0.client [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.253881] env[65788]: WARNING openstack [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.254401] env[65788]: WARNING openstack [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.265565] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e59be32b-b62c-4b78-9aef-d87f6218dc9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.282438] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created folder: Project (69c8fb9f477e45068cd7c1e4defb64a0) in parent group-v910111. [ 1074.282723] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating folder: Instances. Parent ref: group-v910401. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1074.282972] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae5e2fd3-8798-4637-b8ac-c302df204f2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.298163] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created folder: Instances in parent group-v910401. [ 1074.298474] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1074.298731] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1074.298964] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea0873ac-ea81-448c-9831-1d092fac1a1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.315594] env[65788]: DEBUG nova.scheduler.client.report [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1074.327947] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1074.327947] env[65788]: value = "task-4663029" [ 1074.327947] env[65788]: _type = "Task" [ 1074.327947] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.344413] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663029, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.355014] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.445144] env[65788]: WARNING openstack [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.445144] env[65788]: WARNING openstack [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.465887] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d62226-4111-fcd2-181c-57c4574c4798, 'name': SearchDatastore_Task, 'duration_secs': 0.090964} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.465887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1074.465887] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7/1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.465887] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f7cc6357-f0f8-46ff-b5f1-71ac59f3f185 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.473792] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1074.473792] env[65788]: value = "task-4663030" [ 1074.473792] env[65788]: _type = "Task" [ 1074.473792] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.488260] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663030, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.525764] env[65788]: WARNING neutronclient.v2_0.client [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1074.525895] env[65788]: WARNING openstack [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1074.526247] env[65788]: WARNING openstack [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1074.634082] env[65788]: DEBUG nova.network.neutron [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updated VIF entry in instance network info cache for port e7eece2d-f24b-4fcf-b1d8-a232af2c0021. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1074.634518] env[65788]: DEBUG nova.network.neutron [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1074.821778] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.822986] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1074.827837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.525s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.829712] env[65788]: INFO nova.compute.claims [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.847205] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663029, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.860160] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663026, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.566358} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.860524] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef/OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef.vmdk to [datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk. [ 1074.860723] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Cleaning up location [datastore2] OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1074.860899] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_c8841fb8-083d-4fdc-92cb-5b18ade20eef {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.861259] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88cb2dee-eca9-4a9e-af6b-71363c280971 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.872078] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1074.872078] env[65788]: value = "task-4663031" [ 1074.872078] env[65788]: _type = "Task" [ 1074.872078] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.884242] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.986060] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663030, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.005717] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Successfully updated port: 2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1075.137526] env[65788]: DEBUG oslo_concurrency.lockutils [req-0a59a7ff-7274-49c2-b241-638d9d81f0da req-bdda5845-0f4e-409c-a4fa-0ae3823ad8e0 service nova] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1075.335211] env[65788]: DEBUG nova.compute.utils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1075.338722] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1075.338939] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1075.339278] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.339573] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.340170] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1075.340516] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1075.360237] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663029, 'name': CreateVM_Task, 'duration_secs': 0.690554} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.361084] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1075.361470] env[65788]: WARNING neutronclient.v2_0.client [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1075.361593] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.361770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.362099] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1075.362376] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd9a4d13-4b3d-45bc-86de-fd658c8867f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.368141] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1075.368141] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be9e7b-200f-294a-c04c-04e00238257b" [ 1075.368141] env[65788]: _type = "Task" [ 1075.368141] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.381534] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be9e7b-200f-294a-c04c-04e00238257b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.388234] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.388583] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.388791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1075.389134] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk to [datastore2] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1075.389842] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e66cd49-c0e6-43d4-a95b-d071a643a6b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.402187] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1075.402187] env[65788]: value = "task-4663032" [ 1075.402187] env[65788]: _type = "Task" [ 1075.402187] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.411976] env[65788]: DEBUG nova.policy [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd335bd92f4f647c6b7ec8678b11b5613', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d8d51f6df554bcba9bd0e8dbbbcc3d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1075.421685] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.484122] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663030, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544686} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.484476] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7/1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1075.484752] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.485039] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-645fb525-e7e6-4567-a303-2f2f97bb916c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.496098] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1075.496098] env[65788]: value = "task-4663033" [ 1075.496098] env[65788]: _type = "Task" [ 1075.496098] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.507073] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.509024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "refresh_cache-c5dfbb9b-430d-4e93-b24d-e918d90e123e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.509024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "refresh_cache-c5dfbb9b-430d-4e93-b24d-e918d90e123e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.509234] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1075.689890] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166326e8-fd5c-4c8b-be7c-fd5ea7459034 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.699671] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebcc17f-bd30-4dc1-8139-15d3f52ad882 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.734262] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Successfully created port: c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1075.737283] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b061579-155c-4699-a33b-3a664ee56ce7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.747174] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c9d0b1-9625-40e9-881f-712a00c542c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.763684] env[65788]: DEBUG nova.compute.provider_tree [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.839553] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1075.880599] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be9e7b-200f-294a-c04c-04e00238257b, 'name': SearchDatastore_Task, 'duration_secs': 0.010757} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.880940] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1075.881208] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1075.881488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.881846] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1075.882091] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1075.883191] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-149b33ea-dbf8-4829-9e35-76043b100316 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.902319] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1075.902522] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1075.903379] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d26c6dd6-8fcb-4aae-85dc-193877b412f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.914551] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1075.914551] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52677104-f2ee-619f-0e34-aedb64090825" [ 1075.914551] env[65788]: _type = "Task" [ 1075.914551] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.918188] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663032, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.923670] env[65788]: DEBUG nova.compute.manager [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Received event network-vif-plugged-2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1075.923947] env[65788]: DEBUG oslo_concurrency.lockutils [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Acquiring lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1075.924222] env[65788]: DEBUG oslo_concurrency.lockutils [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1075.924443] env[65788]: DEBUG oslo_concurrency.lockutils [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.924662] env[65788]: DEBUG nova.compute.manager [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] No waiting events found dispatching network-vif-plugged-2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1075.924897] env[65788]: WARNING nova.compute.manager [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Received unexpected event network-vif-plugged-2b9f08a0-53d5-41f1-9780-2a30d0c7d529 for instance with vm_state building and task_state spawning. [ 1075.925241] env[65788]: DEBUG nova.compute.manager [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Received event network-changed-2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1075.925453] env[65788]: DEBUG nova.compute.manager [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Refreshing instance network info cache due to event network-changed-2b9f08a0-53d5-41f1-9780-2a30d0c7d529. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1075.925672] env[65788]: DEBUG oslo_concurrency.lockutils [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Acquiring lock "refresh_cache-c5dfbb9b-430d-4e93-b24d-e918d90e123e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.934645] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52677104-f2ee-619f-0e34-aedb64090825, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.008830] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154472} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.009296] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1076.011079] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748babaf-d31b-4dbd-b5a8-d918cb55d041 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.014919] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.015383] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.049219] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7/1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.049700] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a70f455-d576-41e1-bb30-2241e0be5c16 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.067716] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1076.079660] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1076.079660] env[65788]: value = "task-4663034" [ 1076.079660] env[65788]: _type = "Task" [ 1076.079660] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.091922] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663034, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.093787] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.094310] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.177893] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.178860] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.179333] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.267711] env[65788]: DEBUG nova.scheduler.client.report [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1076.342247] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Updating instance_info_cache with network_info: [{"id": "2b9f08a0-53d5-41f1-9780-2a30d0c7d529", "address": "fa:16:3e:4e:71:be", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b9f08a0-53", "ovs_interfaceid": "2b9f08a0-53d5-41f1-9780-2a30d0c7d529", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1076.417678] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663032, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.431596] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52677104-f2ee-619f-0e34-aedb64090825, 'name': SearchDatastore_Task, 'duration_secs': 0.097473} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.432516] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5321629-6631-44b5-a031-be12f4b14dec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.441107] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1076.441107] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e1a408-e893-68fa-1c89-2061a9722305" [ 1076.441107] env[65788]: _type = "Task" [ 1076.441107] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.452297] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e1a408-e893-68fa-1c89-2061a9722305, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.597949] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.773984] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.946s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.774612] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1076.777902] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.583s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.779420] env[65788]: INFO nova.compute.claims [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.848570] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "refresh_cache-c5dfbb9b-430d-4e93-b24d-e918d90e123e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.849014] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Instance network_info: |[{"id": "2b9f08a0-53d5-41f1-9780-2a30d0c7d529", "address": "fa:16:3e:4e:71:be", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b9f08a0-53", "ovs_interfaceid": "2b9f08a0-53d5-41f1-9780-2a30d0c7d529", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1076.850124] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1076.852662] env[65788]: DEBUG oslo_concurrency.lockutils [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Acquired lock "refresh_cache-c5dfbb9b-430d-4e93-b24d-e918d90e123e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.852849] env[65788]: DEBUG nova.network.neutron [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Refreshing network info cache for port 2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1076.854788] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:71:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b9f08a0-53d5-41f1-9780-2a30d0c7d529', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1076.864949] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Creating folder: Project (5d8d51f6df554bcba9bd0e8dbbbcc3d4). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1076.866860] env[65788]: WARNING neutronclient.v2_0.client [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1076.867651] env[65788]: WARNING openstack [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1076.868068] env[65788]: WARNING openstack [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1076.875504] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4add1d8-96ba-4be4-8590-dc414cd1416c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.887525] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1076.887816] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1076.887994] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1076.888200] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1076.888351] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1076.888503] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1076.888719] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.888879] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1076.889086] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1076.889301] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1076.889476] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1076.890816] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c926282-e112-4831-9991-4988113614ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.895240] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Created folder: Project (5d8d51f6df554bcba9bd0e8dbbbcc3d4) in parent group-v910111. [ 1076.895446] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Creating folder: Instances. Parent ref: group-v910404. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1076.896235] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5812f6ee-337a-450c-bdec-fe66b70ba431 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.902605] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7ec5f5-737f-4bdf-ace0-7958679385d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.911999] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Created folder: Instances in parent group-v910404. [ 1076.912316] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1076.913013] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1076.921794] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49733347-3493-439b-8dac-fad7eed9a336 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.942223] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663032, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.948446] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1076.948446] env[65788]: value = "task-4663037" [ 1076.948446] env[65788]: _type = "Task" [ 1076.948446] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.956112] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e1a408-e893-68fa-1c89-2061a9722305, 'name': SearchDatastore_Task, 'duration_secs': 0.091369} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.956835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1076.957158] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e/475291c8-a16d-4980-9ef4-7ed6a4dc382e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1076.957454] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08c98667-0e48-421f-88b5-ceb20f14bc57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.963275] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663037, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.972207] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1076.972207] env[65788]: value = "task-4663038" [ 1076.972207] env[65788]: _type = "Task" [ 1076.972207] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.983910] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.033966] env[65788]: WARNING openstack [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.034547] env[65788]: WARNING openstack [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.094647] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.162342] env[65788]: WARNING neutronclient.v2_0.client [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.163360] env[65788]: WARNING openstack [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.163870] env[65788]: WARNING openstack [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.241533] env[65788]: DEBUG nova.compute.manager [req-f1013d86-64a3-46d8-8cec-67387631ed4e req-d56707d2-11f4-4d0c-81ef-221ce0a6936c service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Received event network-vif-plugged-c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1077.241959] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1013d86-64a3-46d8-8cec-67387631ed4e req-d56707d2-11f4-4d0c-81ef-221ce0a6936c service nova] Acquiring lock "8755c56a-0f5e-4e95-a8b8-c643849fa798-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1077.242238] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1013d86-64a3-46d8-8cec-67387631ed4e req-d56707d2-11f4-4d0c-81ef-221ce0a6936c service nova] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1077.242510] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1013d86-64a3-46d8-8cec-67387631ed4e req-d56707d2-11f4-4d0c-81ef-221ce0a6936c service nova] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1077.242773] env[65788]: DEBUG nova.compute.manager [req-f1013d86-64a3-46d8-8cec-67387631ed4e req-d56707d2-11f4-4d0c-81ef-221ce0a6936c service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] No waiting events found dispatching network-vif-plugged-c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1077.243198] env[65788]: WARNING nova.compute.manager [req-f1013d86-64a3-46d8-8cec-67387631ed4e req-d56707d2-11f4-4d0c-81ef-221ce0a6936c service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Received unexpected event network-vif-plugged-c26e3ef9-b46e-47db-8ad1-682fc4c31a62 for instance with vm_state building and task_state spawning. [ 1077.264758] env[65788]: DEBUG nova.network.neutron [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Updated VIF entry in instance network info cache for port 2b9f08a0-53d5-41f1-9780-2a30d0c7d529. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1077.265177] env[65788]: DEBUG nova.network.neutron [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Updating instance_info_cache with network_info: [{"id": "2b9f08a0-53d5-41f1-9780-2a30d0c7d529", "address": "fa:16:3e:4e:71:be", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b9f08a0-53", "ovs_interfaceid": "2b9f08a0-53d5-41f1-9780-2a30d0c7d529", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1077.284340] env[65788]: DEBUG nova.compute.utils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1077.288104] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1077.288524] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1077.288810] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.289146] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1077.289741] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1077.290109] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1077.342540] env[65788]: DEBUG nova.policy [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd335bd92f4f647c6b7ec8678b11b5613', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d8d51f6df554bcba9bd0e8dbbbcc3d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1077.348050] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Successfully updated port: c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1077.424315] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663032, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.462507] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663037, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.490088] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.598024] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.650629] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84aeffc-1527-4866-a8dd-93abb119f4fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.663046] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Successfully created port: ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1077.667909] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e1c4f7-bb9f-455f-ade4-edf0c73f4d2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.710047] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e47f8d2-a0a5-41ed-9f64-d41ef506eace {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.722260] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90a2cd3-33fb-4436-8cff-9cc179c636d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.741132] env[65788]: DEBUG nova.compute.provider_tree [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.768536] env[65788]: DEBUG oslo_concurrency.lockutils [req-57d29064-3c76-465b-bbcd-16eb187679db req-299fb63f-12c3-4047-814d-c03f70be73af service nova] Releasing lock "refresh_cache-c5dfbb9b-430d-4e93-b24d-e918d90e123e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1077.789438] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1077.851803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "refresh_cache-8755c56a-0f5e-4e95-a8b8-c643849fa798" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.852009] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "refresh_cache-8755c56a-0f5e-4e95-a8b8-c643849fa798" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1077.852202] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1077.920778] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663032, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.504346} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.920778] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/17403798-a0dc-480e-ac18-3ae8959fda51/17403798-a0dc-480e-ac18-3ae8959fda51.vmdk to [datastore2] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1077.921897] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4274b6-1019-42c0-8c4e-974f87b402ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.950866] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1077.951277] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dd2868e-0400-430a-87df-12054266d138 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.975561] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663037, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.977350] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1077.977350] env[65788]: value = "task-4663039" [ 1077.977350] env[65788]: _type = "Task" [ 1077.977350] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.989940] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663038, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.993807] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663039, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.094392] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.246627] env[65788]: DEBUG nova.scheduler.client.report [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.357311] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.357744] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.397213] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1078.418015] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.418439] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.466411] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663037, 'name': CreateVM_Task, 'duration_secs': 1.414946} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.466411] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1078.466818] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.466961] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.467137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.467467] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1078.467737] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cea8109-2af5-46ce-8a93-22a861387841 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.474354] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1078.474354] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5224a321-e232-c284-243f-6f27ba160039" [ 1078.474354] env[65788]: _type = "Task" [ 1078.474354] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.485465] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1078.486231] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1078.486508] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1078.497940] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5224a321-e232-c284-243f-6f27ba160039, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.501663] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.501890] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.502139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.502311] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1078.502466] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.502753] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663038, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.454343} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.502962] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b37c8163-aa8b-443f-8f43-37b2230ad6af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.504916] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e/475291c8-a16d-4980-9ef4-7ed6a4dc382e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1078.505149] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1078.508748] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-276e3768-5bee-442d-954b-e99e935eb145 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.510824] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663039, 'name': ReconfigVM_Task, 'duration_secs': 0.357828} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.511613] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 4dcf8335-701d-4386-9de4-f14f5d375d1f/4dcf8335-701d-4386-9de4-f14f5d375d1f.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.512765] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_name': '/dev/sda', 'size': 0, 'boot_index': 0, 'disk_bus': None, 'device_type': 'disk', 'encrypted': False, 'encryption_secret_uuid': None, 'guest_format': None, 'encryption_options': None, 'encryption_format': None, 'image_id': 'd5a7b45a-a1f3-4d64-bb88-d3ab995bc18b'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '6d58f422-5a00-4162-b476-54f8f15b4bc8', 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910389', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'name': 'volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '4dcf8335-701d-4386-9de4-f14f5d375d1f', 'attached_at': '', 'detached_at': '', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'serial': '7fc1ca8c-4392-42f3-9430-fec6de904bb6'}, 'disk_bus': None, 'device_type': None, 'delete_on_termination': False, 'mount_device': '/dev/sdb', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=65788) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1078.512973] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1078.513189] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910389', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'name': 'volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '4dcf8335-701d-4386-9de4-f14f5d375d1f', 'attached_at': '', 'detached_at': '', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'serial': '7fc1ca8c-4392-42f3-9430-fec6de904bb6'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1078.514535] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f7812c-efec-4525-99bd-6c37c5a6d375 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.519835] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1078.519835] env[65788]: value = "task-4663040" [ 1078.519835] env[65788]: _type = "Task" [ 1078.519835] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.521293] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.521490] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1078.541187] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6361220-bca7-4e1b-806c-d2cc757aba02 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.544578] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a9ddf0-fae5-4197-b97e-196e9d478599 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.556836] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663040, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.572309] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1078.572309] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287b644-8b72-4a06-307b-3cbdc54319b8" [ 1078.572309] env[65788]: _type = "Task" [ 1078.572309] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.580412] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6/volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.580850] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f315604-9220-493f-af87-a27809cad11b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.598996] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Updating instance_info_cache with network_info: [{"id": "c26e3ef9-b46e-47db-8ad1-682fc4c31a62", "address": "fa:16:3e:c3:c9:ae", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc26e3ef9-b4", "ovs_interfaceid": "c26e3ef9-b46e-47db-8ad1-682fc4c31a62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1078.613873] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287b644-8b72-4a06-307b-3cbdc54319b8, 'name': SearchDatastore_Task, 'duration_secs': 0.010952} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.619239] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663034, 'name': ReconfigVM_Task, 'duration_secs': 2.114203} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.619554] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1078.619554] env[65788]: value = "task-4663041" [ 1078.619554] env[65788]: _type = "Task" [ 1078.619554] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.620503] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a781bd3a-e709-47e4-8fbc-54faa81a1c93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.622922] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7/1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.623960] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64933d7b-5745-46ae-a428-0de4369fd7c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.633341] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1078.633341] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5234626f-38de-d456-04c3-8cd1cf2e062d" [ 1078.633341] env[65788]: _type = "Task" [ 1078.633341] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.637558] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.642473] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1078.642473] env[65788]: value = "task-4663042" [ 1078.642473] env[65788]: _type = "Task" [ 1078.642473] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.650792] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5234626f-38de-d456-04c3-8cd1cf2e062d, 'name': SearchDatastore_Task, 'duration_secs': 0.012531} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.651624] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.651934] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] c5dfbb9b-430d-4e93-b24d-e918d90e123e/c5dfbb9b-430d-4e93-b24d-e918d90e123e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1078.652263] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94b26709-436b-42d2-904b-02bfbfeba727 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.658480] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663042, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.667028] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1078.667028] env[65788]: value = "task-4663043" [ 1078.667028] env[65788]: _type = "Task" [ 1078.667028] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.677596] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663043, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.754355] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.976s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.755176] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1078.758028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.800s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1078.758220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.760706] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.699s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1078.762335] env[65788]: INFO nova.compute.claims [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.800368] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1078.806652] env[65788]: INFO nova.scheduler.client.report [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Deleted allocations for instance d8103991-fd1f-4a57-81ce-1a47dc4defe0 [ 1078.841178] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1078.841603] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1078.841879] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1078.845281] env[65788]: DEBUG nova.virt.hardware [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1078.846078] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d194cd-1d08-44e3-b44a-26476082478d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.859349] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9b83e8-1128-4d4b-a5a3-07975825ff95 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.039212] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663040, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075999} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.039212] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.039212] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcd5dca-cf46-4e64-8061-3e09239ccf4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.067106] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e/475291c8-a16d-4980-9ef4-7ed6a4dc382e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.067532] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b542c35d-1772-4347-bb06-23bd2c1bf054 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.090791] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1079.090791] env[65788]: value = "task-4663044" [ 1079.090791] env[65788]: _type = "Task" [ 1079.090791] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.102485] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663044, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.105263] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "refresh_cache-8755c56a-0f5e-4e95-a8b8-c643849fa798" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.105936] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Instance network_info: |[{"id": "c26e3ef9-b46e-47db-8ad1-682fc4c31a62", "address": "fa:16:3e:c3:c9:ae", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc26e3ef9-b4", "ovs_interfaceid": "c26e3ef9-b46e-47db-8ad1-682fc4c31a62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1079.106259] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:c9:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c26e3ef9-b46e-47db-8ad1-682fc4c31a62', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1079.114963] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1079.115287] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1079.115546] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-053bc7b7-f120-40de-a7b2-cbb5ad458d6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.143203] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663041, 'name': ReconfigVM_Task, 'duration_secs': 0.385584} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.148562] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfigured VM instance instance-0000004b to attach disk [datastore2] volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6/volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.153785] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1079.153785] env[65788]: value = "task-4663045" [ 1079.153785] env[65788]: _type = "Task" [ 1079.153785] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.154248] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b51b125f-ac92-49e7-8706-809b251a286f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.176643] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663042, 'name': Rename_Task, 'duration_secs': 0.177869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.177312] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1079.177847] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e88d91fa-b7d5-4eb8-83ef-799ebcb06b3f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.183252] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663045, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.187551] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663043, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.187846] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1079.187846] env[65788]: value = "task-4663046" [ 1079.187846] env[65788]: _type = "Task" [ 1079.187846] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.189080] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1079.189080] env[65788]: value = "task-4663047" [ 1079.189080] env[65788]: _type = "Task" [ 1079.189080] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.201136] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663047, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.204695] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663046, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.267749] env[65788]: DEBUG nova.compute.utils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1079.271862] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1079.272084] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1079.272427] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.272770] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.273399] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1079.273858] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1079.318170] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a1fe5911-a678-4b19-9f19-a5a35483331e tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "d8103991-fd1f-4a57-81ce-1a47dc4defe0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.241s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.334166] env[65788]: DEBUG nova.policy [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72684a9d23904034af7dcda5026dcf93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcd247e1f9bb4cc5b29e009f727bacca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1079.576432] env[65788]: DEBUG nova.compute.manager [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Received event network-changed-c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1079.576778] env[65788]: DEBUG nova.compute.manager [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Refreshing instance network info cache due to event network-changed-c26e3ef9-b46e-47db-8ad1-682fc4c31a62. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1079.577180] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Acquiring lock "refresh_cache-8755c56a-0f5e-4e95-a8b8-c643849fa798" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.577180] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Acquired lock "refresh_cache-8755c56a-0f5e-4e95-a8b8-c643849fa798" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.577323] env[65788]: DEBUG nova.network.neutron [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Refreshing network info cache for port c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1079.603849] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663044, 'name': ReconfigVM_Task, 'duration_secs': 0.51368} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.604017] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e/475291c8-a16d-4980-9ef4-7ed6a4dc382e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.604818] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8e404941-f1c3-4869-81ed-1c3167961c47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.612399] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1079.612399] env[65788]: value = "task-4663048" [ 1079.612399] env[65788]: _type = "Task" [ 1079.612399] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.624291] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663048, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.679167] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663043, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.682377] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663045, 'name': CreateVM_Task, 'duration_secs': 0.440761} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.682662] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1079.683182] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1079.683572] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.683744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.684323] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1079.684392] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1cd5e0a-4e3d-4bc9-a9f1-666dce834922 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.687443] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Successfully created port: 647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1079.694209] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1079.694209] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5266f0c5-4f3f-d41a-f603-d3e98b91a46b" [ 1079.694209] env[65788]: _type = "Task" [ 1079.694209] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.697223] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Successfully updated port: ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1079.717776] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663046, 'name': ReconfigVM_Task, 'duration_secs': 0.28075} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.719031] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910389', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'name': 'volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '4dcf8335-701d-4386-9de4-f14f5d375d1f', 'attached_at': '', 'detached_at': '', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'serial': '7fc1ca8c-4392-42f3-9430-fec6de904bb6'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1079.719914] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c066fad2-f236-4458-92e9-38442801b4c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.733853] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663047, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.734166] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5266f0c5-4f3f-d41a-f603-d3e98b91a46b, 'name': SearchDatastore_Task, 'duration_secs': 0.010748} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.734893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.735151] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.735400] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.735551] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.735763] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.736015] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d4c7ef0-c688-478d-b489-b41aef23b20a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.740683] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1079.740683] env[65788]: value = "task-4663049" [ 1079.740683] env[65788]: _type = "Task" [ 1079.740683] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.746990] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.747340] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1079.751843] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-336be46a-bfcc-4cb5-ba91-e505c67f9fc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.754735] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663049, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.759726] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1079.759726] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dbed80-0435-ceee-5ff0-2d3df7760877" [ 1079.759726] env[65788]: _type = "Task" [ 1079.759726] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.772195] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dbed80-0435-ceee-5ff0-2d3df7760877, 'name': SearchDatastore_Task, 'duration_secs': 0.010395} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.773073] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c36b937-3481-42d6-8753-2cf44f799dda {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.780599] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1079.780599] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528b4ff5-c6a0-4c12-8ecc-4a62048eda36" [ 1079.780599] env[65788]: _type = "Task" [ 1079.780599] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.787035] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1079.796857] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528b4ff5-c6a0-4c12-8ecc-4a62048eda36, 'name': SearchDatastore_Task, 'duration_secs': 0.011192} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.796857] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1079.796989] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 8755c56a-0f5e-4e95-a8b8-c643849fa798/8755c56a-0f5e-4e95-a8b8-c643849fa798.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1079.797186] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfdbe492-f64b-4b65-aa49-f2435894a473 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.806221] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1079.806221] env[65788]: value = "task-4663050" [ 1079.806221] env[65788]: _type = "Task" [ 1079.806221] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.814443] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663050, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.074130] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd6b968-7b29-4637-bad1-9e91a32cc675 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.079878] env[65788]: WARNING neutronclient.v2_0.client [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.080604] env[65788]: WARNING openstack [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.080998] env[65788]: WARNING openstack [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.091849] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cd2615-d127-46f9-aa8b-b5f1795f6dfc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.133074] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa24a21f-5362-4d83-ad7f-780c840b3d98 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.143792] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663048, 'name': Rename_Task, 'duration_secs': 0.17832} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.146358] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.146695] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7cd8d2e-995b-44c5-8a6b-9a5f02367adc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.149429] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4287e276-126e-4f9a-8883-2b2c9368b2e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.166011] env[65788]: DEBUG nova.compute.provider_tree [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.169181] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1080.169181] env[65788]: value = "task-4663051" [ 1080.169181] env[65788]: _type = "Task" [ 1080.169181] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.183129] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663051, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.187770] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663043, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.505046} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.188115] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] c5dfbb9b-430d-4e93-b24d-e918d90e123e/c5dfbb9b-430d-4e93-b24d-e918d90e123e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1080.188356] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.188665] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-237308a1-61f9-4a6c-8ca1-1ee75a8e372b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.196616] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1080.196616] env[65788]: value = "task-4663052" [ 1080.196616] env[65788]: _type = "Task" [ 1080.196616] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.214751] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663052, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.215446] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "refresh_cache-398791b2-9dc0-478a-8ed8-bdfbace0404e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.215575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "refresh_cache-398791b2-9dc0-478a-8ed8-bdfbace0404e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1080.215747] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1080.218628] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663047, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.254640] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663049, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.319713] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663050, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.445101] env[65788]: WARNING openstack [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.445516] env[65788]: WARNING openstack [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.672047] env[65788]: DEBUG nova.scheduler.client.report [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1080.685945] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663051, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.690487] env[65788]: WARNING neutronclient.v2_0.client [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1080.691170] env[65788]: WARNING openstack [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.691526] env[65788]: WARNING openstack [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.715057] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663047, 'name': PowerOnVM_Task} progress is 71%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.718495] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663052, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079041} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.719541] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1080.719883] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1080.727201] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1080.728407] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b848c1d9-c600-468b-b69b-41ae155be8a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.755129] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] c5dfbb9b-430d-4e93-b24d-e918d90e123e/c5dfbb9b-430d-4e93-b24d-e918d90e123e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.759447] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dba57d47-b5ae-4723-88d3-605564243620 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.785897] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663049, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.787512] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1080.787512] env[65788]: value = "task-4663053" [ 1080.787512] env[65788]: _type = "Task" [ 1080.787512] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.797131] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663053, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.800506] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1080.817768] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663050, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.909833} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.822549] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 8755c56a-0f5e-4e95-a8b8-c643849fa798/8755c56a-0f5e-4e95-a8b8-c643849fa798.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1080.822807] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.825949] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aef23b17-d091-42c3-b1a2-e28bef26f7e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.836085] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1080.836085] env[65788]: value = "task-4663054" [ 1080.836085] env[65788]: _type = "Task" [ 1080.836085] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.839091] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1080.839400] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1080.839681] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1080.839881] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1080.840117] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1080.840294] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1080.840512] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.840672] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1080.840876] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1080.841063] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1080.841271] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1080.842374] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1298087-93f3-4ec4-b5b0-7e8ebfee58ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.846131] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1080.860764] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663054, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.862273] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d307aa-778a-463c-b390-63e7119cbb86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.002112] env[65788]: DEBUG nova.network.neutron [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Updated VIF entry in instance network info cache for port c26e3ef9-b46e-47db-8ad1-682fc4c31a62. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1081.002726] env[65788]: DEBUG nova.network.neutron [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Updating instance_info_cache with network_info: [{"id": "c26e3ef9-b46e-47db-8ad1-682fc4c31a62", "address": "fa:16:3e:c3:c9:ae", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc26e3ef9-b4", "ovs_interfaceid": "c26e3ef9-b46e-47db-8ad1-682fc4c31a62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1081.047014] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.047595] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.179512] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.180071] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1081.182701] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.259s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.182916] env[65788]: DEBUG nova.objects.instance [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lazy-loading 'resources' on Instance uuid a35ced42-4317-49b4-b4cc-4ed7e2c85c64 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.190294] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663051, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.210720] env[65788]: DEBUG oslo_vmware.api [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663047, 'name': PowerOnVM_Task, 'duration_secs': 1.780335} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.210926] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.211112] env[65788]: INFO nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Took 12.23 seconds to spawn the instance on the hypervisor. [ 1081.211301] env[65788]: DEBUG nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1081.212127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b531b6-9789-4dca-ab46-8f45b3f43301 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.256578] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663049, 'name': Rename_Task, 'duration_secs': 1.353748} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.256895] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1081.257316] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5e73d8b-b164-4f79-b575-b349c24c3bae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.268139] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1081.268139] env[65788]: value = "task-4663055" [ 1081.268139] env[65788]: _type = "Task" [ 1081.268139] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.269276] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Successfully updated port: 647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1081.282666] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.300173] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663053, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.356299] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663054, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127008} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.356709] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.357779] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcd7a22-0e8c-4758-8a10-4c9d2c0a281b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.383489] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 8755c56a-0f5e-4e95-a8b8-c643849fa798/8755c56a-0f5e-4e95-a8b8-c643849fa798.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.383915] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a11e7445-f2ca-4f0f-83da-7f6db5c635c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.406262] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1081.406262] env[65788]: value = "task-4663056" [ 1081.406262] env[65788]: _type = "Task" [ 1081.406262] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.416846] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.442381] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.443196] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.443597] env[65788]: WARNING openstack [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.507441] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f5eb001-c7e4-49af-b788-5580f68eee18 req-eb546153-0ff8-4c48-927c-253251b066ac service nova] Releasing lock "refresh_cache-8755c56a-0f5e-4e95-a8b8-c643849fa798" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1081.599028] env[65788]: DEBUG nova.network.neutron [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Updating instance_info_cache with network_info: [{"id": "ef95980e-7eaa-4733-ae0d-445fe80490ea", "address": "fa:16:3e:81:c0:18", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef95980e-7e", "ovs_interfaceid": "ef95980e-7eaa-4733-ae0d-445fe80490ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1081.688289] env[65788]: DEBUG nova.compute.utils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1081.692280] env[65788]: DEBUG oslo_vmware.api [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663051, 'name': PowerOnVM_Task, 'duration_secs': 1.404226} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.693097] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1081.693366] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1081.693708] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.694121] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1081.695045] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1081.695199] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1081.703807] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.704059] env[65788]: INFO nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Took 9.83 seconds to spawn the instance on the hypervisor. [ 1081.704254] env[65788]: DEBUG nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1081.706562] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63b9161-992b-4b5d-b3d5-b879ddab1d7d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.714255] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Received event network-vif-plugged-ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1081.714537] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Acquiring lock "398791b2-9dc0-478a-8ed8-bdfbace0404e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.714890] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.715109] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.715283] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] No waiting events found dispatching network-vif-plugged-ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1081.715495] env[65788]: WARNING nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Received unexpected event network-vif-plugged-ef95980e-7eaa-4733-ae0d-445fe80490ea for instance with vm_state building and task_state spawning. [ 1081.715670] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Received event network-changed-ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1081.715739] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Refreshing instance network info cache due to event network-changed-ef95980e-7eaa-4733-ae0d-445fe80490ea. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1081.715898] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Acquiring lock "refresh_cache-398791b2-9dc0-478a-8ed8-bdfbace0404e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.731845] env[65788]: INFO nova.compute.manager [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Took 35.75 seconds to build instance. [ 1081.754333] env[65788]: DEBUG nova.policy [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72684a9d23904034af7dcda5026dcf93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcd247e1f9bb4cc5b29e009f727bacca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1081.781535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "refresh_cache-2d74b823-6e28-444d-a80b-c91c2d595460" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.781797] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "refresh_cache-2d74b823-6e28-444d-a80b-c91c2d595460" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1081.781984] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1081.792774] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663055, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.808507] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663053, 'name': ReconfigVM_Task, 'duration_secs': 0.590047} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.809393] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Reconfigured VM instance instance-00000062 to attach disk [datastore2] c5dfbb9b-430d-4e93-b24d-e918d90e123e/c5dfbb9b-430d-4e93-b24d-e918d90e123e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.809859] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c292ee4-93fd-4f6f-81b6-8afbbf3049e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.818753] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1081.818753] env[65788]: value = "task-4663057" [ 1081.818753] env[65788]: _type = "Task" [ 1081.818753] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.830526] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663057, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.928462] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663056, 'name': ReconfigVM_Task, 'duration_secs': 0.305517} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.928462] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 8755c56a-0f5e-4e95-a8b8-c643849fa798/8755c56a-0f5e-4e95-a8b8-c643849fa798.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.928462] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37599550-d6bd-4cde-9141-d1ae28c38f96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.933462] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1081.933462] env[65788]: value = "task-4663058" [ 1081.933462] env[65788]: _type = "Task" [ 1081.933462] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.948118] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663058, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.040738] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628b89f1-31a0-490a-bf0e-20b8fe0a01ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.051370] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766a0f2a-aef4-43ce-8581-9a44fa8d0b0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.088414] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dbb3a7-f6e3-4247-82ff-0a10cb868de9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.094488] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Successfully created port: 4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1082.102257] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e0c96c-c0b8-4bf0-9209-4bce7aa7b1a0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.107850] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "refresh_cache-398791b2-9dc0-478a-8ed8-bdfbace0404e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.108422] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Instance network_info: |[{"id": "ef95980e-7eaa-4733-ae0d-445fe80490ea", "address": "fa:16:3e:81:c0:18", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef95980e-7e", "ovs_interfaceid": "ef95980e-7eaa-4733-ae0d-445fe80490ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1082.108759] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Acquired lock "refresh_cache-398791b2-9dc0-478a-8ed8-bdfbace0404e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.109112] env[65788]: DEBUG nova.network.neutron [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Refreshing network info cache for port ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1082.110703] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:c0:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f762954-6ca5-4da5-bf0a-5d31c51ec570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef95980e-7eaa-4733-ae0d-445fe80490ea', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.119794] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1082.124035] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1082.124035] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08b9ae5f-de9d-46af-980e-fd2df88d9838 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.148229] env[65788]: DEBUG nova.compute.provider_tree [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.156877] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.156877] env[65788]: value = "task-4663059" [ 1082.156877] env[65788]: _type = "Task" [ 1082.156877] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.167260] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663059, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.194100] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1082.233442] env[65788]: INFO nova.compute.manager [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Took 33.14 seconds to build instance. [ 1082.233809] env[65788]: DEBUG oslo_concurrency.lockutils [None req-66f33a34-b6dd-4538-a327-3cdde5d15cd2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.673s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.284850] env[65788]: DEBUG oslo_vmware.api [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663055, 'name': PowerOnVM_Task, 'duration_secs': 0.930171} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.284850] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.286527] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.296799] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1082.330741] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663057, 'name': Rename_Task, 'duration_secs': 0.204729} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.331437] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.332026] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dcc4c540-4981-41f0-9af8-c08b220867e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.340299] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1082.340299] env[65788]: value = "task-4663060" [ 1082.340299] env[65788]: _type = "Task" [ 1082.340299] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.350458] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.412564] env[65788]: DEBUG nova.compute.manager [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1082.413688] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102df4b6-08e1-4395-b1d2-83e589d21b28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.445716] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663058, 'name': Rename_Task, 'duration_secs': 0.17619} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.446045] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.446554] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b26d80e4-4c27-4323-b43f-6cf3491cd717 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.456853] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1082.456853] env[65788]: value = "task-4663061" [ 1082.456853] env[65788]: _type = "Task" [ 1082.456853] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.466473] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.623330] env[65788]: WARNING neutronclient.v2_0.client [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1082.624365] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.624860] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.641299] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1082.652663] env[65788]: DEBUG nova.scheduler.client.report [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1082.668573] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663059, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.736511] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dceb1314-8664-4f28-b55a-39d2b8f9a3dc tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.650s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.753958] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.754397] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.856065] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663060, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.937220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd28a89e-adcf-4d0f-adcf-9a1eb4b2a940 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 47.236s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.950514] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1082.951217] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1082.971239] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663061, 'name': PowerOnVM_Task, 'duration_secs': 0.512912} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.971586] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1082.971735] env[65788]: INFO nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Took 6.12 seconds to spawn the instance on the hypervisor. [ 1082.972649] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1082.973107] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b325f7a-904e-47b3-b35c-111cf1506657 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.011835] env[65788]: DEBUG nova.compute.manager [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-changed-1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1083.011926] env[65788]: DEBUG nova.compute.manager [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Refreshing instance network info cache due to event network-changed-1426fd1f-0507-498d-9a05-9bce1efd861a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1083.012170] env[65788]: DEBUG oslo_concurrency.lockutils [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Acquiring lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.012320] env[65788]: DEBUG oslo_concurrency.lockutils [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Acquired lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.016064] env[65788]: DEBUG nova.network.neutron [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Refreshing network info cache for port 1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1083.018900] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.019614] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.020326] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.158124] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.975s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.161318] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.097s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.161763] env[65788]: DEBUG nova.objects.instance [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'resources' on Instance uuid 64acef8e-640d-4195-9519-ce31d984ddde {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.166349] env[65788]: WARNING neutronclient.v2_0.client [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.167170] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.168854] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.189232] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663059, 'name': CreateVM_Task, 'duration_secs': 0.549591} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.190346] env[65788]: INFO nova.scheduler.client.report [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Deleted allocations for instance a35ced42-4317-49b4-b4cc-4ed7e2c85c64 [ 1083.191857] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.194507] env[65788]: WARNING neutronclient.v2_0.client [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.194990] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.195087] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.195405] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1083.198821] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10dddcf2-1909-4f89-a7ae-e7b78c993b08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.208253] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1083.211038] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1083.211038] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5268d19e-a9a9-4f54-f257-f095fcead404" [ 1083.211038] env[65788]: _type = "Task" [ 1083.211038] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.222194] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5268d19e-a9a9-4f54-f257-f095fcead404, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.242482] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1083.242736] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1083.242844] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1083.243178] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1083.243297] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1083.243583] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1083.243783] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1083.244076] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1083.244321] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1083.244595] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1083.244818] env[65788]: DEBUG nova.virt.hardware [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1083.245766] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2817093f-6af6-4e35-bf77-57f604fe1d61 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.255409] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db972434-e4b0-4b61-9ce0-544bfd4ad2bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.315076] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Updating instance_info_cache with network_info: [{"id": "647f361e-2e88-4fc5-a2d1-d713f32cdf72", "address": "fa:16:3e:7d:5f:cc", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap647f361e-2e", "ovs_interfaceid": "647f361e-2e88-4fc5-a2d1-d713f32cdf72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1083.360719] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.360971] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.362474] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663060, 'name': PowerOnVM_Task, 'duration_secs': 0.753566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.363116] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.363327] env[65788]: INFO nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Took 9.44 seconds to spawn the instance on the hypervisor. [ 1083.363507] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1083.364390] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98bb66f8-2d48-41b3-bb76-0a7626c343f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.443893] env[65788]: DEBUG nova.network.neutron [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Updated VIF entry in instance network info cache for port ef95980e-7eaa-4733-ae0d-445fe80490ea. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1083.444292] env[65788]: DEBUG nova.network.neutron [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Updating instance_info_cache with network_info: [{"id": "ef95980e-7eaa-4733-ae0d-445fe80490ea", "address": "fa:16:3e:81:c0:18", "network": {"id": "595ef2b4-5862-4246-8c66-3ef06367982a", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-262464530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d8d51f6df554bcba9bd0e8dbbbcc3d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f762954-6ca5-4da5-bf0a-5d31c51ec570", "external-id": "nsx-vlan-transportzone-930", "segmentation_id": 930, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef95980e-7e", "ovs_interfaceid": "ef95980e-7eaa-4733-ae0d-445fe80490ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1083.492956] env[65788]: INFO nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Took 33.24 seconds to build instance. [ 1083.533209] env[65788]: WARNING neutronclient.v2_0.client [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1083.534681] env[65788]: WARNING openstack [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1083.534969] env[65788]: WARNING openstack [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1083.710072] env[65788]: DEBUG oslo_concurrency.lockutils [None req-512a0a8d-2b83-493c-ab78-07b3b55f0b92 tempest-ServersNegativeTestJSON-460036087 tempest-ServersNegativeTestJSON-460036087-project-member] Lock "a35ced42-4317-49b4-b4cc-4ed7e2c85c64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.229s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.725021] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5268d19e-a9a9-4f54-f257-f095fcead404, 'name': SearchDatastore_Task, 'duration_secs': 0.014161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.725021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.725021] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.725021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.725021] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.725021] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.728356] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a3cebbd-89ef-401a-8ad5-6b544929da5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.740024] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.740024] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.740024] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e07eed9-6d4a-4804-9e5c-17fcd75c2ba5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.752135] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1083.752135] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b5fbd6-795f-a5e6-8333-b656a22be01e" [ 1083.752135] env[65788]: _type = "Task" [ 1083.752135] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.767474] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b5fbd6-795f-a5e6-8333-b656a22be01e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.769090] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Successfully updated port: 4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1083.823662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "refresh_cache-2d74b823-6e28-444d-a80b-c91c2d595460" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.824173] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Instance network_info: |[{"id": "647f361e-2e88-4fc5-a2d1-d713f32cdf72", "address": "fa:16:3e:7d:5f:cc", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap647f361e-2e", "ovs_interfaceid": "647f361e-2e88-4fc5-a2d1-d713f32cdf72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1083.825625] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:5f:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '647f361e-2e88-4fc5-a2d1-d713f32cdf72', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.835678] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1083.839153] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1083.840064] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa6f9983-c15c-4376-9d5b-adbccdcc0a7e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.863967] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1083.868334] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.868334] env[65788]: value = "task-4663062" [ 1083.868334] env[65788]: _type = "Task" [ 1083.868334] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.897029] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663062, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.898472] env[65788]: INFO nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Took 33.68 seconds to build instance. [ 1083.947233] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Releasing lock "refresh_cache-398791b2-9dc0-478a-8ed8-bdfbace0404e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.947550] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Received event network-vif-plugged-647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1083.947788] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Acquiring lock "2d74b823-6e28-444d-a80b-c91c2d595460-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1083.948054] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Lock "2d74b823-6e28-444d-a80b-c91c2d595460-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1083.948246] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Lock "2d74b823-6e28-444d-a80b-c91c2d595460-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.948511] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] No waiting events found dispatching network-vif-plugged-647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1083.948754] env[65788]: WARNING nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Received unexpected event network-vif-plugged-647f361e-2e88-4fc5-a2d1-d713f32cdf72 for instance with vm_state building and task_state spawning. [ 1083.948969] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Received event network-changed-647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1083.949231] env[65788]: DEBUG nova.compute.manager [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Refreshing instance network info cache due to event network-changed-647f361e-2e88-4fc5-a2d1-d713f32cdf72. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1083.949601] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Acquiring lock "refresh_cache-2d74b823-6e28-444d-a80b-c91c2d595460" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.949792] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Acquired lock "refresh_cache-2d74b823-6e28-444d-a80b-c91c2d595460" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1083.949980] env[65788]: DEBUG nova.network.neutron [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Refreshing network info cache for port 647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1083.996702] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.772s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.021607] env[65788]: WARNING openstack [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.022036] env[65788]: WARNING openstack [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.080474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba323a4-90b3-4daa-8a88-6a628bbafe8f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.089681] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1c8a85-2d8c-49f7-bca0-538af1563614 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.125372] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49d807b-cc98-468e-95a4-9b0a7482df92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.134810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda27eaa-7fc2-43bf-9350-3cf4e02000f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.153903] env[65788]: DEBUG nova.compute.manager [req-b9199c29-4a81-4d33-aa66-661a13f9e7d8 req-dfe4da29-45e6-48a8-88e4-6a802b7b3c1e service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Received event network-vif-plugged-4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1084.154031] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9199c29-4a81-4d33-aa66-661a13f9e7d8 req-dfe4da29-45e6-48a8-88e4-6a802b7b3c1e service nova] Acquiring lock "a022cb81-5ff8-41aa-9307-c99a58416e94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.154387] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9199c29-4a81-4d33-aa66-661a13f9e7d8 req-dfe4da29-45e6-48a8-88e4-6a802b7b3c1e service nova] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.155138] env[65788]: DEBUG oslo_concurrency.lockutils [req-b9199c29-4a81-4d33-aa66-661a13f9e7d8 req-dfe4da29-45e6-48a8-88e4-6a802b7b3c1e service nova] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.155138] env[65788]: DEBUG nova.compute.manager [req-b9199c29-4a81-4d33-aa66-661a13f9e7d8 req-dfe4da29-45e6-48a8-88e4-6a802b7b3c1e service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] No waiting events found dispatching network-vif-plugged-4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1084.155281] env[65788]: WARNING nova.compute.manager [req-b9199c29-4a81-4d33-aa66-661a13f9e7d8 req-dfe4da29-45e6-48a8-88e4-6a802b7b3c1e service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Received unexpected event network-vif-plugged-4d3fb85b-0714-4b63-ae7c-59527a920bad for instance with vm_state building and task_state spawning. [ 1084.166438] env[65788]: DEBUG nova.compute.provider_tree [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.266839] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b5fbd6-795f-a5e6-8333-b656a22be01e, 'name': SearchDatastore_Task, 'duration_secs': 0.021841} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.267750] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17c59880-2405-4b39-a528-bd8e94332719 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.271456] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "refresh_cache-a022cb81-5ff8-41aa-9307-c99a58416e94" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.271651] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "refresh_cache-a022cb81-5ff8-41aa-9307-c99a58416e94" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.271827] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1084.275247] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1084.275247] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5277a335-4a23-e291-518a-7c16253aea45" [ 1084.275247] env[65788]: _type = "Task" [ 1084.275247] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.287942] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5277a335-4a23-e291-518a-7c16253aea45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.356430] env[65788]: WARNING neutronclient.v2_0.client [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.357154] env[65788]: WARNING openstack [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.357497] env[65788]: WARNING openstack [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.384554] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663062, 'name': CreateVM_Task, 'duration_secs': 0.456846} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.385383] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1084.385383] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.385667] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.385819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.386182] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1084.386457] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5747385-c2c7-4f42-abb4-f5ad38a3042e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.393103] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1084.393103] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5227e1a6-cc47-b098-322b-ffb4656ed3d3" [ 1084.393103] env[65788]: _type = "Task" [ 1084.393103] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.400208] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.400683] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.187s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.404461] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5227e1a6-cc47-b098-322b-ffb4656ed3d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.453026] env[65788]: WARNING neutronclient.v2_0.client [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1084.453884] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.454388] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.670510] env[65788]: DEBUG nova.scheduler.client.report [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1084.778199] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1084.778199] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1084.794931] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5277a335-4a23-e291-518a-7c16253aea45, 'name': SearchDatastore_Task, 'duration_secs': 0.013827} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.794931] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.795628] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 398791b2-9dc0-478a-8ed8-bdfbace0404e/398791b2-9dc0-478a-8ed8-bdfbace0404e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.795628] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-965fd6e1-a78d-4eff-82d7-b70d96cc0611 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.808190] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1084.808190] env[65788]: value = "task-4663063" [ 1084.808190] env[65788]: _type = "Task" [ 1084.808190] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.818568] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.846847] env[65788]: DEBUG nova.network.neutron [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updated VIF entry in instance network info cache for port 1426fd1f-0507-498d-9a05-9bce1efd861a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1084.847298] env[65788]: DEBUG nova.network.neutron [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1084.909806] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5227e1a6-cc47-b098-322b-ffb4656ed3d3, 'name': SearchDatastore_Task, 'duration_secs': 0.013276} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.911715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.911715] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.911715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.911715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1084.911715] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.911715] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bdc7b7ac-bccd-45a9-a930-9cf2fdee1b1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.915479] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1084.921424] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.921621] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1084.923449] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d14606d-b0eb-47a9-b3c4-13cad6abb9dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.930081] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1084.930081] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521216fd-0ee9-6ac7-8077-6cfe3723f85f" [ 1084.930081] env[65788]: _type = "Task" [ 1084.930081] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.943950] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521216fd-0ee9-6ac7-8077-6cfe3723f85f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.006044] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.006097] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.027786] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.028216] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.126913] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1085.127796] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.128790] env[65788]: WARNING openstack [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.177687] env[65788]: WARNING neutronclient.v2_0.client [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1085.179474] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1085.180428] env[65788]: WARNING openstack [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1085.192210] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.031s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.196441] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.142s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.228395] env[65788]: INFO nova.scheduler.client.report [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocations for instance 64acef8e-640d-4195-9519-ce31d984ddde [ 1085.322644] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663063, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.356210] env[65788]: DEBUG oslo_concurrency.lockutils [req-628e3393-dd0e-4756-b2ee-937fbac1f19b req-ff9dd611-73ba-449a-9824-0148706394df service nova] Releasing lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.401398] env[65788]: DEBUG nova.network.neutron [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Updating instance_info_cache with network_info: [{"id": "4d3fb85b-0714-4b63-ae7c-59527a920bad", "address": "fa:16:3e:7b:74:23", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d3fb85b-07", "ovs_interfaceid": "4d3fb85b-0714-4b63-ae7c-59527a920bad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.415254] env[65788]: DEBUG nova.network.neutron [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Updated VIF entry in instance network info cache for port 647f361e-2e88-4fc5-a2d1-d713f32cdf72. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1085.415254] env[65788]: DEBUG nova.network.neutron [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Updating instance_info_cache with network_info: [{"id": "647f361e-2e88-4fc5-a2d1-d713f32cdf72", "address": "fa:16:3e:7d:5f:cc", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap647f361e-2e", "ovs_interfaceid": "647f361e-2e88-4fc5-a2d1-d713f32cdf72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1085.450199] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521216fd-0ee9-6ac7-8077-6cfe3723f85f, 'name': SearchDatastore_Task, 'duration_secs': 0.013178} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.450199] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba2467a-ebd0-4fd5-9e86-a2071fd1d32b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.460936] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1085.460936] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f2d6ee-fee8-c254-0e98-c542579bd847" [ 1085.460936] env[65788]: _type = "Task" [ 1085.460936] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.474877] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f2d6ee-fee8-c254-0e98-c542579bd847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.703364] env[65788]: INFO nova.compute.claims [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.744606] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fef25f7-0c18-497b-89d8-774ef3b6b1f2 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "64acef8e-640d-4195-9519-ce31d984ddde" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.808s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.825076] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.63901} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.825430] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 398791b2-9dc0-478a-8ed8-bdfbace0404e/398791b2-9dc0-478a-8ed8-bdfbace0404e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.825640] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.825985] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fa7a6ba-3134-41e3-9ca9-37864549d3e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.834218] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1085.834218] env[65788]: value = "task-4663064" [ 1085.834218] env[65788]: _type = "Task" [ 1085.834218] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.845812] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663064, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.904909] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "refresh_cache-a022cb81-5ff8-41aa-9307-c99a58416e94" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.905479] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Instance network_info: |[{"id": "4d3fb85b-0714-4b63-ae7c-59527a920bad", "address": "fa:16:3e:7b:74:23", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d3fb85b-07", "ovs_interfaceid": "4d3fb85b-0714-4b63-ae7c-59527a920bad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1085.905958] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:74:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f6fb0104-186b-4288-b87e-634893f46f01', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d3fb85b-0714-4b63-ae7c-59527a920bad', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.915347] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1085.915660] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1085.915952] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c977205-5b76-4ecb-a12a-1a513803052a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.932687] env[65788]: DEBUG oslo_concurrency.lockutils [req-d138d584-ac93-4641-8602-746673a640c3 req-e72ac4aa-8b5e-42fc-9561-e0795ab23bf4 service nova] Releasing lock "refresh_cache-2d74b823-6e28-444d-a80b-c91c2d595460" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.939320] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1085.939320] env[65788]: value = "task-4663065" [ 1085.939320] env[65788]: _type = "Task" [ 1085.939320] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.951294] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663065, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.972631] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f2d6ee-fee8-c254-0e98-c542579bd847, 'name': SearchDatastore_Task, 'duration_secs': 0.02326} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.973013] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1085.973335] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 2d74b823-6e28-444d-a80b-c91c2d595460/2d74b823-6e28-444d-a80b-c91c2d595460.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1085.973644] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae028609-9e83-4bee-bedc-9666a6366ad2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.981089] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1085.981089] env[65788]: value = "task-4663066" [ 1085.981089] env[65788]: _type = "Task" [ 1085.981089] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.989324] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.210110] env[65788]: INFO nova.compute.resource_tracker [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating resource usage from migration b2330f68-7793-465a-8fed-06f3654996c8 [ 1086.215923] env[65788]: DEBUG nova.compute.manager [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Received event network-changed-4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1086.216149] env[65788]: DEBUG nova.compute.manager [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Refreshing instance network info cache due to event network-changed-4d3fb85b-0714-4b63-ae7c-59527a920bad. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1086.216438] env[65788]: DEBUG oslo_concurrency.lockutils [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Acquiring lock "refresh_cache-a022cb81-5ff8-41aa-9307-c99a58416e94" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.216515] env[65788]: DEBUG oslo_concurrency.lockutils [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Acquired lock "refresh_cache-a022cb81-5ff8-41aa-9307-c99a58416e94" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.216679] env[65788]: DEBUG nova.network.neutron [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Refreshing network info cache for port 4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1086.347811] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663064, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095523} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.348380] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.349157] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5169728c-3082-4414-b05d-132f56c1f76f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.377444] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 398791b2-9dc0-478a-8ed8-bdfbace0404e/398791b2-9dc0-478a-8ed8-bdfbace0404e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.380632] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf90f752-144f-4c65-84f3-4d3568ad02b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.405553] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1086.405553] env[65788]: value = "task-4663067" [ 1086.405553] env[65788]: _type = "Task" [ 1086.405553] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.418343] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663067, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.451790] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663065, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.491534] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663066, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.543258] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48438dc1-dd06-4e88-96a5-7ce231d43c0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.555185] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76af3ebb-4f4e-4366-a7ec-064d6d638f54 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.592239] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bbb020-91a9-4670-9234-fc08323caf14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.603362] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a137b4-4ca8-43af-831e-8438e80f8924 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.620673] env[65788]: DEBUG nova.compute.provider_tree [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1086.719618] env[65788]: WARNING neutronclient.v2_0.client [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.720432] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.720777] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.819594] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1086.820049] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1086.918276] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663067, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.952197] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663065, 'name': CreateVM_Task, 'duration_secs': 0.836488} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.952366] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1086.952895] env[65788]: WARNING neutronclient.v2_0.client [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1086.953301] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.953458] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1086.953834] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1086.954142] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3c1f290-1aef-40a1-9cf3-87720300dee9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.959480] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1086.959480] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520f93ca-fe36-9eba-2ff2-98a5b0b1191e" [ 1086.959480] env[65788]: _type = "Task" [ 1086.959480] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.967991] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520f93ca-fe36-9eba-2ff2-98a5b0b1191e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.991596] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663066, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.782722} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.991851] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 2d74b823-6e28-444d-a80b-c91c2d595460/2d74b823-6e28-444d-a80b-c91c2d595460.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1086.992082] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1086.992358] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a4dae12-9a59-41cc-aaf5-89fca3698d42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.000880] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1087.000880] env[65788]: value = "task-4663068" [ 1087.000880] env[65788]: _type = "Task" [ 1087.000880] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.009262] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663068, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.150682] env[65788]: ERROR nova.scheduler.client.report [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [req-beb41bb8-2827-4f71-981f-272ce3617dd1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-beb41bb8-2827-4f71-981f-272ce3617dd1"}]} [ 1087.168164] env[65788]: DEBUG nova.scheduler.client.report [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1087.185579] env[65788]: DEBUG nova.scheduler.client.report [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1087.185829] env[65788]: DEBUG nova.compute.provider_tree [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1087.206077] env[65788]: DEBUG nova.scheduler.client.report [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1087.213065] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.213273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.236521] env[65788]: DEBUG nova.scheduler.client.report [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1087.315652] env[65788]: WARNING neutronclient.v2_0.client [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1087.316404] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1087.318782] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1087.415220] env[65788]: DEBUG nova.network.neutron [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Updated VIF entry in instance network info cache for port 4d3fb85b-0714-4b63-ae7c-59527a920bad. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1087.415593] env[65788]: DEBUG nova.network.neutron [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Updating instance_info_cache with network_info: [{"id": "4d3fb85b-0714-4b63-ae7c-59527a920bad", "address": "fa:16:3e:7b:74:23", "network": {"id": "8565958b-3de1-4cad-a6a9-db4b5793eabd", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1646636861-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcd247e1f9bb4cc5b29e009f727bacca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f6fb0104-186b-4288-b87e-634893f46f01", "external-id": "nsx-vlan-transportzone-73", "segmentation_id": 73, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d3fb85b-07", "ovs_interfaceid": "4d3fb85b-0714-4b63-ae7c-59527a920bad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1087.423254] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663067, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.472912] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520f93ca-fe36-9eba-2ff2-98a5b0b1191e, 'name': SearchDatastore_Task, 'duration_secs': 0.051867} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.475716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.475977] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.476238] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.476390] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.476605] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.477036] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4773b7af-5baa-4e03-9122-30dc0e26da9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.492473] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.492665] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1087.493476] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a82d7533-b4f4-4ec4-b3a7-63c8ad5a756f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.501200] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1087.501200] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d0c102-ef5e-be66-68cb-622e98ba4c75" [ 1087.501200] env[65788]: _type = "Task" [ 1087.501200] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.512259] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d0c102-ef5e-be66-68cb-622e98ba4c75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.517403] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663068, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.185875} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.517837] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1087.518608] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16752a1e-c402-4ed9-9340-0e10d38f54aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.545029] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 2d74b823-6e28-444d-a80b-c91c2d595460/2d74b823-6e28-444d-a80b-c91c2d595460.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.548077] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11b7185f-3cc4-4aa8-82e7-87dcb1176e1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.570057] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1087.570057] env[65788]: value = "task-4663069" [ 1087.570057] env[65788]: _type = "Task" [ 1087.570057] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.574951] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd27e5fa-a581-43e0-b24e-f8ab4a31f8e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.588458] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eaee0d4-3e55-4d9d-94b1-720e88509955 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.591914] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663069, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.621816] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf64d55-aacc-4e3e-b0ac-dd58b5f0e545 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.630632] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d43edca-7d68-45ba-a865-b1266dde7c0d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.646994] env[65788]: DEBUG nova.compute.provider_tree [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1087.718722] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1087.918391] env[65788]: DEBUG oslo_concurrency.lockutils [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Releasing lock "refresh_cache-a022cb81-5ff8-41aa-9307-c99a58416e94" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1087.918713] env[65788]: DEBUG nova.compute.manager [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Received event network-changed-e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1087.918807] env[65788]: DEBUG nova.compute.manager [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Refreshing instance network info cache due to event network-changed-e7eece2d-f24b-4fcf-b1d8-a232af2c0021. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1087.919016] env[65788]: DEBUG oslo_concurrency.lockutils [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.919192] env[65788]: DEBUG oslo_concurrency.lockutils [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.919359] env[65788]: DEBUG nova.network.neutron [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Refreshing network info cache for port e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1087.920587] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663067, 'name': ReconfigVM_Task, 'duration_secs': 1.373287} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.921056] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 398791b2-9dc0-478a-8ed8-bdfbace0404e/398791b2-9dc0-478a-8ed8-bdfbace0404e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.921718] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66711cbb-f108-4d41-a970-5f1ebf11d9f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.929486] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1087.929486] env[65788]: value = "task-4663070" [ 1087.929486] env[65788]: _type = "Task" [ 1087.929486] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.939049] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663070, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.013057] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d0c102-ef5e-be66-68cb-622e98ba4c75, 'name': SearchDatastore_Task, 'duration_secs': 0.057663} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.013942] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76b3d71f-7725-4b51-986b-00f2254838b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.020322] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1088.020322] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a31e35-126b-69ba-4e21-638ed635f832" [ 1088.020322] env[65788]: _type = "Task" [ 1088.020322] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.031701] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a31e35-126b-69ba-4e21-638ed635f832, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.080841] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663069, 'name': ReconfigVM_Task, 'duration_secs': 0.452721} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.081164] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 2d74b823-6e28-444d-a80b-c91c2d595460/2d74b823-6e28-444d-a80b-c91c2d595460.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.082099] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-228234fd-c1b1-4a9c-be89-b608ba52c448 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.090067] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1088.090067] env[65788]: value = "task-4663071" [ 1088.090067] env[65788]: _type = "Task" [ 1088.090067] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.099648] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663071, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.185625] env[65788]: DEBUG nova.scheduler.client.report [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 140 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1088.186044] env[65788]: DEBUG nova.compute.provider_tree [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 140 to 141 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1088.186241] env[65788]: DEBUG nova.compute.provider_tree [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1088.238569] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.423226] env[65788]: WARNING neutronclient.v2_0.client [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.423226] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.423722] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.444856] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663070, 'name': Rename_Task, 'duration_secs': 0.191666} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.445166] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.445431] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2a14e7b-8320-463b-a48f-0d4122b1a469 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.453660] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1088.453660] env[65788]: value = "task-4663072" [ 1088.453660] env[65788]: _type = "Task" [ 1088.453660] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.464794] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663072, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.537440] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a31e35-126b-69ba-4e21-638ed635f832, 'name': SearchDatastore_Task, 'duration_secs': 0.018872} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.537713] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.537982] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] a022cb81-5ff8-41aa-9307-c99a58416e94/a022cb81-5ff8-41aa-9307-c99a58416e94.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1088.538306] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fa3f9855-954a-4fda-a62a-caba948d3f8d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.547350] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1088.547350] env[65788]: value = "task-4663073" [ 1088.547350] env[65788]: _type = "Task" [ 1088.547350] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.557520] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663073, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.567756] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.568202] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.605410] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663071, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.632310] env[65788]: WARNING neutronclient.v2_0.client [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1088.633105] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1088.633479] env[65788]: WARNING openstack [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1088.692873] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.496s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.693126] env[65788]: INFO nova.compute.manager [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Migrating [ 1088.701378] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.898s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.701878] env[65788]: DEBUG nova.objects.instance [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lazy-loading 'resources' on Instance uuid 31a6981e-89cd-4b83-85cc-36dd163e1f8f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.724140] env[65788]: DEBUG nova.network.neutron [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updated VIF entry in instance network info cache for port e7eece2d-f24b-4fcf-b1d8-a232af2c0021. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1088.724549] env[65788]: DEBUG nova.network.neutron [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1088.967289] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663072, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.060025] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663073, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.102846] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663071, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.214031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.214410] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1089.214706] env[65788]: DEBUG nova.network.neutron [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1089.227745] env[65788]: DEBUG oslo_concurrency.lockutils [req-878ac8a8-1768-4a57-a138-61aa99fd54d9 req-6305284a-3f5f-49e3-a30a-8b474a9dfd9f service nova] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1089.468625] env[65788]: DEBUG oslo_vmware.api [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663072, 'name': PowerOnVM_Task, 'duration_secs': 0.861264} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.469056] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.469297] env[65788]: INFO nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Took 10.67 seconds to spawn the instance on the hypervisor. [ 1089.469921] env[65788]: DEBUG nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1089.470383] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcae3fba-5b46-4bf0-b5e4-5a846072890e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.506354] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df5094f-51a2-438d-a2e8-e1044a9b868a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.517192] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9803bfbc-9953-47a5-9aa6-7565172bc95a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.549262] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56d9fca-1a0a-4a9e-a1b9-407832580aa7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.561245] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663073, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548047} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.563507] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] a022cb81-5ff8-41aa-9307-c99a58416e94/a022cb81-5ff8-41aa-9307-c99a58416e94.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1089.563817] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.564148] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7f10543-7217-4444-943f-bc75bfbba9cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.567197] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f2f6fb-485e-48d6-99e3-6649ff9e223f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.582480] env[65788]: DEBUG nova.compute.provider_tree [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.586465] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1089.586465] env[65788]: value = "task-4663074" [ 1089.586465] env[65788]: _type = "Task" [ 1089.586465] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.602016] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.608063] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663071, 'name': Rename_Task, 'duration_secs': 1.142005} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.608376] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.608662] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf5d3231-7dcc-4c10-a145-4b2429e6b1ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.616515] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1089.616515] env[65788]: value = "task-4663075" [ 1089.616515] env[65788]: _type = "Task" [ 1089.616515] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.625943] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663075, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.720505] env[65788]: WARNING neutronclient.v2_0.client [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.721431] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1089.721887] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1089.860815] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1089.860936] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1089.935141] env[65788]: WARNING neutronclient.v2_0.client [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1089.935850] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1089.936225] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1089.997499] env[65788]: INFO nova.compute.manager [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Took 39.71 seconds to build instance. [ 1090.034583] env[65788]: DEBUG nova.network.neutron [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1090.087460] env[65788]: DEBUG nova.scheduler.client.report [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1090.103136] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072406} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.104515] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1090.105767] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5af91ba-63f1-416c-be48-aee0919497fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.131548] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] a022cb81-5ff8-41aa-9307-c99a58416e94/a022cb81-5ff8-41aa-9307-c99a58416e94.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1090.135461] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8629a657-9fcb-42b8-9417-25ec79c0ff4c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.160640] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663075, 'name': PowerOnVM_Task} progress is 78%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.161809] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1090.161809] env[65788]: value = "task-4663076" [ 1090.161809] env[65788]: _type = "Task" [ 1090.161809] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.171532] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663076, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.499993] env[65788]: DEBUG oslo_concurrency.lockutils [None req-63904635-3e9a-4795-a139-76e566af587d tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.234s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.537837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1090.596527] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.895s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.601332] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.931s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.604593] env[65788]: INFO nova.compute.claims [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.630307] env[65788]: INFO nova.scheduler.client.report [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Deleted allocations for instance 31a6981e-89cd-4b83-85cc-36dd163e1f8f [ 1090.635878] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663075, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.675704] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663076, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.710922] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.710922] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.711118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.711304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.711490] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.713745] env[65788]: INFO nova.compute.manager [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Terminating instance [ 1091.139027] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663075, 'name': PowerOnVM_Task, 'duration_secs': 1.053937} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.139027] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1091.139027] env[65788]: INFO nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Took 10.34 seconds to spawn the instance on the hypervisor. [ 1091.139027] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1091.139027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb84122-22a3-4dac-8f18-c2acbcb2145a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.146012] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9b0118a6-925e-4d7e-ae28-51e99864d234 tempest-ServerTagsTestJSON-2143161080 tempest-ServerTagsTestJSON-2143161080-project-member] Lock "31a6981e-89cd-4b83-85cc-36dd163e1f8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.886s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1091.175357] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663076, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.219419] env[65788]: DEBUG nova.compute.manager [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1091.219419] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.221801] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9979dc07-44b6-40f0-aa5f-31c1d9c3f1f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.231912] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.232433] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8ea13c6-175b-491f-ab3d-424ad63ea2d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.242564] env[65788]: DEBUG oslo_vmware.api [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1091.242564] env[65788]: value = "task-4663077" [ 1091.242564] env[65788]: _type = "Task" [ 1091.242564] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.253486] env[65788]: DEBUG oslo_vmware.api [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663077, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.661056] env[65788]: INFO nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Took 40.49 seconds to build instance. [ 1091.674971] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663076, 'name': ReconfigVM_Task, 'duration_secs': 1.246502} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.677969] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Reconfigured VM instance instance-00000066 to attach disk [datastore1] a022cb81-5ff8-41aa-9307-c99a58416e94/a022cb81-5ff8-41aa-9307-c99a58416e94.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1091.679574] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63190b88-938b-4896-9efa-fd7dbcd6e9da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.687860] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1091.687860] env[65788]: value = "task-4663078" [ 1091.687860] env[65788]: _type = "Task" [ 1091.687860] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.702121] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663078, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.759509] env[65788]: DEBUG oslo_vmware.api [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663077, 'name': PowerOffVM_Task, 'duration_secs': 0.49653} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.759886] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.760082] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.760358] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1be14463-1a8d-4c3e-85f4-da1cba39e40f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.843948] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.844340] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.844628] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleting the datastore file [datastore2] c5dfbb9b-430d-4e93-b24d-e918d90e123e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.844938] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-16686027-dc75-46cc-8406-0d464ea0690b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.853115] env[65788]: DEBUG oslo_vmware.api [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1091.853115] env[65788]: value = "task-4663080" [ 1091.853115] env[65788]: _type = "Task" [ 1091.853115] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.865799] env[65788]: DEBUG oslo_vmware.api [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.908128] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b470a08b-f460-4122-b1d7-389df57bc753 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.918740] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b81dc6-4ece-4610-bebe-70c9bd5fb1b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.958046] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74433de3-7131-4860-ae44-7ca94c037b43 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.966994] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceeb588b-4a68-411f-bccf-2868ee606308 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.983227] env[65788]: DEBUG nova.compute.provider_tree [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1092.054429] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e177294-fd2e-4f8b-ae47-517577a993ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.074358] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1092.165724] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "2d74b823-6e28-444d-a80b-c91c2d595460" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.002s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.198392] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663078, 'name': Rename_Task, 'duration_secs': 0.264509} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.198969] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1092.199068] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45d42edc-0de3-4f35-b509-b69b0e0e2413 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.206746] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1092.206746] env[65788]: value = "task-4663081" [ 1092.206746] env[65788]: _type = "Task" [ 1092.206746] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.216137] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.369795] env[65788]: DEBUG oslo_vmware.api [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203152} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.370087] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.370279] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.370447] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.373283] env[65788]: INFO nova.compute.manager [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1092.373283] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1092.373283] env[65788]: DEBUG nova.compute.manager [-] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1092.373283] env[65788]: DEBUG nova.network.neutron [-] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1092.373283] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.373283] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1092.373283] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1092.415638] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1092.494045] env[65788]: DEBUG nova.scheduler.client.report [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1092.581229] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.581560] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ded0c68-94ce-4c02-b12b-58be7ef4eecf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.590545] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1092.590545] env[65788]: value = "task-4663082" [ 1092.590545] env[65788]: _type = "Task" [ 1092.590545] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.600315] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.673449] env[65788]: DEBUG nova.compute.manager [req-c08c9766-656d-475b-a502-b72060fab94e req-5ecc6b5c-9a2d-4949-a824-9fa294b356dc service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Received event network-vif-deleted-2b9f08a0-53d5-41f1-9780-2a30d0c7d529 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1092.673449] env[65788]: INFO nova.compute.manager [req-c08c9766-656d-475b-a502-b72060fab94e req-5ecc6b5c-9a2d-4949-a824-9fa294b356dc service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Neutron deleted interface 2b9f08a0-53d5-41f1-9780-2a30d0c7d529; detaching it from the instance and deleting it from the info cache [ 1092.673449] env[65788]: DEBUG nova.network.neutron [req-c08c9766-656d-475b-a502-b72060fab94e req-5ecc6b5c-9a2d-4949-a824-9fa294b356dc service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1092.720176] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663081, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.998662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.999425] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1093.006055] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.606s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.011027] env[65788]: INFO nova.compute.claims [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.101931] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663082, 'name': PowerOffVM_Task, 'duration_secs': 0.391839} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.102244] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.102437] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1093.133974] env[65788]: DEBUG nova.network.neutron [-] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1093.179302] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5bd040a7-1c52-46cb-ab93-93ef8e29a25f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.191483] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cbb0c2-21be-4a66-be60-bab6c37017c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.217936] env[65788]: DEBUG oslo_vmware.api [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663081, 'name': PowerOnVM_Task, 'duration_secs': 0.8572} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.234181] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1093.234450] env[65788]: INFO nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Took 10.03 seconds to spawn the instance on the hypervisor. [ 1093.234659] env[65788]: DEBUG nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1093.235491] env[65788]: DEBUG nova.compute.manager [req-c08c9766-656d-475b-a502-b72060fab94e req-5ecc6b5c-9a2d-4949-a824-9fa294b356dc service nova] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Detach interface failed, port_id=2b9f08a0-53d5-41f1-9780-2a30d0c7d529, reason: Instance c5dfbb9b-430d-4e93-b24d-e918d90e123e could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1093.236471] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abbaa5e-f897-4252-99d3-78799fa50bef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.522152] env[65788]: DEBUG nova.compute.utils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1093.523676] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1093.523984] env[65788]: DEBUG nova.network.neutron [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1093.524507] env[65788]: WARNING neutronclient.v2_0.client [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.524904] env[65788]: WARNING neutronclient.v2_0.client [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1093.525529] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1093.525880] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1093.580744] env[65788]: DEBUG nova.policy [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2625758e73c64384982cb820ea055cb1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b70d3c8627449eaa6372ebe3bd90233', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1093.609204] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1093.609468] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1093.609659] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1093.609880] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1093.610040] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1093.610186] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1093.610424] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.610592] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1093.610777] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1093.610977] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1093.611686] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1093.617018] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-747ae05f-0fc1-4440-ba75-ed4b8185cba9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.637150] env[65788]: INFO nova.compute.manager [-] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Took 1.26 seconds to deallocate network for instance. [ 1093.681789] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1093.681789] env[65788]: value = "task-4663083" [ 1093.681789] env[65788]: _type = "Task" [ 1093.681789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.755556] env[65788]: INFO nova.compute.manager [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Took 37.72 seconds to build instance. [ 1093.931891] env[65788]: DEBUG nova.network.neutron [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Successfully created port: dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1094.033765] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1094.144245] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.195089] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663083, 'name': ReconfigVM_Task, 'duration_secs': 0.476901} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.197836] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1094.257795] env[65788]: DEBUG oslo_concurrency.lockutils [None req-de09ada5-793a-416e-b5e3-dad0a5b833d5 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.045s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.311323] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecb564f-e3b5-42da-9a2f-5b70b5f3a463 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.322196] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfdb574-d988-4f2f-a69f-0bc138a932e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.355361] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123578ec-2c7b-4c9a-a37f-4a7c61428382 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.363987] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444abd41-c08c-40a6-b29e-9353e6029ada {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.380738] env[65788]: DEBUG nova.compute.provider_tree [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.704621] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1094.704886] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1094.705055] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1094.705248] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1094.705398] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1094.705552] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1094.705826] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.706064] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1094.706251] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1094.706417] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1094.706596] env[65788]: DEBUG nova.virt.hardware [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1094.712027] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1094.712919] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0826a211-17c1-49eb-98c1-3c36184d3339 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.733654] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1094.733654] env[65788]: value = "task-4663084" [ 1094.733654] env[65788]: _type = "Task" [ 1094.733654] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.742832] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.883930] env[65788]: DEBUG nova.scheduler.client.report [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1094.887774] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "2d74b823-6e28-444d-a80b-c91c2d595460" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.888138] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "2d74b823-6e28-444d-a80b-c91c2d595460" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.888256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "2d74b823-6e28-444d-a80b-c91c2d595460-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.888451] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "2d74b823-6e28-444d-a80b-c91c2d595460-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.888617] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "2d74b823-6e28-444d-a80b-c91c2d595460-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.890620] env[65788]: INFO nova.compute.manager [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Terminating instance [ 1094.943412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "a022cb81-5ff8-41aa-9307-c99a58416e94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.943664] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.943893] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "a022cb81-5ff8-41aa-9307-c99a58416e94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1094.944104] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1094.944274] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1094.946670] env[65788]: INFO nova.compute.manager [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Terminating instance [ 1095.044717] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1095.072360] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1095.072626] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1095.072785] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1095.072971] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1095.073135] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1095.073283] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1095.073491] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.073775] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1095.074063] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1095.074256] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1095.074442] env[65788]: DEBUG nova.virt.hardware [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1095.075349] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09db8507-53f1-41a5-9384-0afbeed15e14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.084168] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94dfc84-4fc0-4295-8f81-9bebe9ef015f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.244938] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663084, 'name': ReconfigVM_Task, 'duration_secs': 0.178479} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.245361] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1095.246206] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a24ffd-9375-4834-bde9-bd26e615890b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.271014] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-84d81198-187d-46ec-870e-36818081a75e/volume-84d81198-187d-46ec-870e-36818081a75e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.271255] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16ef3c38-f794-40e5-a347-cd7e5b8da1d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.291977] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1095.291977] env[65788]: value = "task-4663085" [ 1095.291977] env[65788]: _type = "Task" [ 1095.291977] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.301844] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663085, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.395024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.395024] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1095.395816] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.157s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.397377] env[65788]: INFO nova.compute.claims [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1095.400067] env[65788]: DEBUG nova.compute.manager [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1095.400466] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.401764] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f997b9a-743e-4cb9-807e-2c13bd48de97 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.410743] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.411248] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72850d67-ec90-45dd-babb-667a7d1d1141 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.420088] env[65788]: DEBUG oslo_vmware.api [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1095.420088] env[65788]: value = "task-4663086" [ 1095.420088] env[65788]: _type = "Task" [ 1095.420088] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.429923] env[65788]: DEBUG oslo_vmware.api [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663086, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.451567] env[65788]: DEBUG nova.compute.manager [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1095.452651] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.453591] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29542ea-f8da-4309-b354-fc62d593b7e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.463459] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.463596] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ec6f60e-ae15-4655-b537-a919fb9d8d29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.469289] env[65788]: DEBUG nova.compute.manager [req-d715d990-e3b8-4a1e-bcff-917a89b5538a req-5a5f89e5-5d12-41e3-8288-d91a91b11028 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-vif-plugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1095.469527] env[65788]: DEBUG oslo_concurrency.lockutils [req-d715d990-e3b8-4a1e-bcff-917a89b5538a req-5a5f89e5-5d12-41e3-8288-d91a91b11028 service nova] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1095.469736] env[65788]: DEBUG oslo_concurrency.lockutils [req-d715d990-e3b8-4a1e-bcff-917a89b5538a req-5a5f89e5-5d12-41e3-8288-d91a91b11028 service nova] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.469905] env[65788]: DEBUG oslo_concurrency.lockutils [req-d715d990-e3b8-4a1e-bcff-917a89b5538a req-5a5f89e5-5d12-41e3-8288-d91a91b11028 service nova] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.470155] env[65788]: DEBUG nova.compute.manager [req-d715d990-e3b8-4a1e-bcff-917a89b5538a req-5a5f89e5-5d12-41e3-8288-d91a91b11028 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] No waiting events found dispatching network-vif-plugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1095.470346] env[65788]: WARNING nova.compute.manager [req-d715d990-e3b8-4a1e-bcff-917a89b5538a req-5a5f89e5-5d12-41e3-8288-d91a91b11028 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received unexpected event network-vif-plugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc for instance with vm_state building and task_state spawning. [ 1095.477397] env[65788]: DEBUG oslo_vmware.api [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1095.477397] env[65788]: value = "task-4663087" [ 1095.477397] env[65788]: _type = "Task" [ 1095.477397] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.488939] env[65788]: DEBUG oslo_vmware.api [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.588546] env[65788]: DEBUG nova.network.neutron [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Successfully updated port: dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1095.806708] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663085, 'name': ReconfigVM_Task, 'duration_secs': 0.361103} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.807212] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-84d81198-187d-46ec-870e-36818081a75e/volume-84d81198-187d-46ec-870e-36818081a75e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.807593] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1095.902753] env[65788]: DEBUG nova.compute.utils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1095.906353] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1095.906440] env[65788]: DEBUG nova.network.neutron [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1095.906761] env[65788]: WARNING neutronclient.v2_0.client [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.907081] env[65788]: WARNING neutronclient.v2_0.client [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1095.908239] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1095.908704] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1095.934552] env[65788]: DEBUG oslo_vmware.api [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663086, 'name': PowerOffVM_Task, 'duration_secs': 0.448898} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.934744] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.934924] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.935221] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1327b21e-04f4-4077-8a2f-746e2ab3278f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.967822] env[65788]: DEBUG nova.policy [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e879429f0504963a38563a287f114f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '021bf6dd87e340969a57e7e5b33268e7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1095.991718] env[65788]: DEBUG oslo_vmware.api [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663087, 'name': PowerOffVM_Task, 'duration_secs': 0.235747} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.991718] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.991718] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.991718] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b2aca1b-3870-4cd5-87d4-9c0fa507b18a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.010540] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1096.011867] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1096.011867] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleting the datastore file [datastore1] 2d74b823-6e28-444d-a80b-c91c2d595460 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.011867] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a04722a7-c3ec-4bfb-8933-39e74506a230 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.019824] env[65788]: DEBUG oslo_vmware.api [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1096.019824] env[65788]: value = "task-4663090" [ 1096.019824] env[65788]: _type = "Task" [ 1096.019824] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.030764] env[65788]: DEBUG oslo_vmware.api [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663090, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.072762] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1096.072762] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1096.072762] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleting the datastore file [datastore1] a022cb81-5ff8-41aa-9307-c99a58416e94 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.073062] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18a1d470-7e7f-4e01-83fd-58dc56ed8edf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.079742] env[65788]: DEBUG oslo_vmware.api [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for the task: (returnval){ [ 1096.079742] env[65788]: value = "task-4663091" [ 1096.079742] env[65788]: _type = "Task" [ 1096.079742] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.088733] env[65788]: DEBUG oslo_vmware.api [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.091429] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.091595] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1096.091789] env[65788]: DEBUG nova.network.neutron [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1096.315224] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36009fdb-3e24-4adf-bac1-44fa229521d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.338398] env[65788]: DEBUG nova.network.neutron [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Successfully created port: 3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1096.341641] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc27b06-eb73-4160-b4c4-c37d97723008 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.362353] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1096.407715] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1096.534114] env[65788]: DEBUG oslo_vmware.api [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663090, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.535186] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.535383] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.535562] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.535738] env[65788]: INFO nova.compute.manager [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1096.536108] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1096.536254] env[65788]: DEBUG nova.compute.manager [-] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1096.536351] env[65788]: DEBUG nova.network.neutron [-] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1096.536624] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.537154] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.537452] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.578934] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.593219] env[65788]: DEBUG oslo_vmware.api [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Task: {'id': task-4663091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238579} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.593350] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.593505] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.593712] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.593953] env[65788]: INFO nova.compute.manager [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1096.594223] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1096.595014] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.595401] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.602423] env[65788]: DEBUG nova.compute.manager [-] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1096.602533] env[65788]: DEBUG nova.network.neutron [-] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1096.602783] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.603233] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.603479] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1096.712086] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30e3515-b254-4f4e-a792-cc16e8dd7699 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.721155] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04ec5f3-50dd-42e7-896f-fbb7912d61b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.754585] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3be54f8-7ea5-4000-af77-f1b0bafda9e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.763043] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e72cef2-bb18-41af-829e-a5b1675e4c09 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.778671] env[65788]: DEBUG nova.compute.provider_tree [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.805373] env[65788]: DEBUG nova.compute.manager [req-1b6509f2-ac2d-497e-99c9-898633fafd85 req-ebe29dcd-3fcd-4459-b856-7a16a9b2681e service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Received event network-vif-deleted-647f361e-2e88-4fc5-a2d1-d713f32cdf72 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1096.805686] env[65788]: INFO nova.compute.manager [req-1b6509f2-ac2d-497e-99c9-898633fafd85 req-ebe29dcd-3fcd-4459-b856-7a16a9b2681e service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Neutron deleted interface 647f361e-2e88-4fc5-a2d1-d713f32cdf72; detaching it from the instance and deleting it from the info cache [ 1096.805802] env[65788]: DEBUG nova.network.neutron [req-1b6509f2-ac2d-497e-99c9-898633fafd85 req-ebe29dcd-3fcd-4459-b856-7a16a9b2681e service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1096.809259] env[65788]: DEBUG nova.network.neutron [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1096.865460] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1096.895428] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1096.895831] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.266914] env[65788]: WARNING neutronclient.v2_0.client [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1097.267639] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1097.267994] env[65788]: WARNING openstack [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1097.282311] env[65788]: DEBUG nova.scheduler.client.report [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1097.286856] env[65788]: DEBUG nova.network.neutron [-] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.308878] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a3196a6-f851-44c1-b8a3-ce902bfda00d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.325433] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f534fab6-9c5d-4622-acc3-21745721432c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.372829] env[65788]: DEBUG nova.compute.manager [req-1b6509f2-ac2d-497e-99c9-898633fafd85 req-ebe29dcd-3fcd-4459-b856-7a16a9b2681e service nova] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Detach interface failed, port_id=647f361e-2e88-4fc5-a2d1-d713f32cdf72, reason: Instance 2d74b823-6e28-444d-a80b-c91c2d595460 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1097.418201] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1097.450289] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1097.450719] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1097.451092] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1097.451424] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1097.451738] env[65788]: DEBUG nova.virt.hardware [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1097.452753] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2dd191-4ea2-446b-8165-76b58ea64b39 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.464878] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6cca86-e3d9-4a1a-af48-66b3d0f71655 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.501762] env[65788]: DEBUG nova.compute.manager [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1097.502140] env[65788]: DEBUG nova.compute.manager [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing instance network info cache due to event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1097.503370] env[65788]: DEBUG oslo_concurrency.lockutils [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.509802] env[65788]: DEBUG nova.network.neutron [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.617337] env[65788]: DEBUG nova.network.neutron [-] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1097.788219] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.788770] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1097.791457] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.648s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.791680] env[65788]: DEBUG nova.objects.instance [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lazy-loading 'resources' on Instance uuid c5dfbb9b-430d-4e93-b24d-e918d90e123e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.792796] env[65788]: INFO nova.compute.manager [-] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Took 1.26 seconds to deallocate network for instance. [ 1097.960297] env[65788]: DEBUG nova.network.neutron [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Successfully updated port: 3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1098.011551] env[65788]: WARNING neutronclient.v2_0.client [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.013906] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.014275] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance network_info: |[{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1098.014805] env[65788]: DEBUG oslo_concurrency.lockutils [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.014985] env[65788]: DEBUG nova.network.neutron [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1098.016255] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:12:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc5727b9-f909-4ed4-80d0-06ba5986f0cc', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1098.023884] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1098.025249] env[65788]: WARNING neutronclient.v2_0.client [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.025867] env[65788]: WARNING openstack [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.026423] env[65788]: WARNING openstack [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.035174] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1098.036415] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab3619b9-f227-4a1a-b984-76b2bb601242 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.052447] env[65788]: DEBUG nova.network.neutron [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Port ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1098.060582] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1098.060582] env[65788]: value = "task-4663092" [ 1098.060582] env[65788]: _type = "Task" [ 1098.060582] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.076369] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663092, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.120673] env[65788]: INFO nova.compute.manager [-] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Took 1.52 seconds to deallocate network for instance. [ 1098.133185] env[65788]: WARNING openstack [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.133918] env[65788]: WARNING openstack [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.220750] env[65788]: WARNING neutronclient.v2_0.client [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.221931] env[65788]: WARNING openstack [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.223697] env[65788]: WARNING openstack [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.298321] env[65788]: DEBUG nova.compute.utils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1098.302166] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.302166] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1098.302166] env[65788]: DEBUG nova.network.neutron [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1098.302874] env[65788]: WARNING neutronclient.v2_0.client [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.302874] env[65788]: WARNING neutronclient.v2_0.client [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.303488] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.303939] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1098.323719] env[65788]: DEBUG nova.network.neutron [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updated VIF entry in instance network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1098.324060] env[65788]: DEBUG nova.network.neutron [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1098.353425] env[65788]: DEBUG nova.policy [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32b4adcc8fa42fba0dbffd4582fcf69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36859c1eb994614b2a77400f811cf50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1098.465831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.465831] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.465831] env[65788]: DEBUG nova.network.neutron [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1098.574970] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663092, 'name': CreateVM_Task, 'duration_secs': 0.364598} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.574970] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1098.575429] env[65788]: WARNING neutronclient.v2_0.client [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1098.575796] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.576019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1098.576363] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1098.576602] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de22e29b-afdb-49e8-a4c3-601877da8271 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.590022] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1098.590022] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523ca492-d5ec-0fd7-ad1a-7aea9e5c86ea" [ 1098.590022] env[65788]: _type = "Task" [ 1098.590022] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.598107] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523ca492-d5ec-0fd7-ad1a-7aea9e5c86ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.600041] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a99f3c4-3dbb-4902-baab-ac2643586434 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.609345] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35eb023-9791-41c1-91e5-5d56abd4b472 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.646782] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1098.648349] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a0ac0d-9515-41b2-ba7c-aab000bfdf56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.660156] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3a1373-32b0-41b3-b5e2-3a4ef8ff7d22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.685632] env[65788]: DEBUG nova.compute.provider_tree [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.718594] env[65788]: DEBUG nova.network.neutron [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Successfully created port: 01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1098.813259] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1098.829884] env[65788]: DEBUG oslo_concurrency.lockutils [req-7fce1825-60c7-4c29-a133-a4ac9f26a2f6 req-706f5018-540f-4828-bede-5d80614bf392 service nova] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.840054] env[65788]: DEBUG nova.compute.manager [req-2f2a7d68-c404-4d88-8344-1ba9fd8ce719 req-40877186-8cb6-4afc-ba5d-bb628ca860ec service nova] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Received event network-vif-deleted-4d3fb85b-0714-4b63-ae7c-59527a920bad {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1098.970554] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1098.970967] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.007438] env[65788]: DEBUG nova.network.neutron [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1099.039804] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.040228] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.081855] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1099.082034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.082096] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.099553] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523ca492-d5ec-0fd7-ad1a-7aea9e5c86ea, 'name': SearchDatastore_Task, 'duration_secs': 0.015194} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.099858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.100089] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1099.100341] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.100485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.100668] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1099.101399] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb3c13a0-3bc1-45bd-8262-480e35596e30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.105279] env[65788]: WARNING neutronclient.v2_0.client [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1099.105921] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1099.106444] env[65788]: WARNING openstack [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1099.116206] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1099.116435] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1099.117235] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecf66b72-c91f-4253-bd74-aaa845d641d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.124462] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1099.124462] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525258e0-bfb4-43b1-7281-22207a8f66fd" [ 1099.124462] env[65788]: _type = "Task" [ 1099.124462] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.134275] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525258e0-bfb4-43b1-7281-22207a8f66fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.190328] env[65788]: DEBUG nova.scheduler.client.report [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1099.210823] env[65788]: DEBUG nova.network.neutron [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updating instance_info_cache with network_info: [{"id": "3ace6194-8918-410f-a12e-0f966ea6e08e", "address": "fa:16:3e:df:f6:3a", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ace6194-89", "ovs_interfaceid": "3ace6194-8918-410f-a12e-0f966ea6e08e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1099.529705] env[65788]: DEBUG nova.compute.manager [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Received event network-vif-plugged-3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1099.529705] env[65788]: DEBUG oslo_concurrency.lockutils [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1099.530288] env[65788]: DEBUG oslo_concurrency.lockutils [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.530288] env[65788]: DEBUG oslo_concurrency.lockutils [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.530552] env[65788]: DEBUG nova.compute.manager [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] No waiting events found dispatching network-vif-plugged-3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1099.530953] env[65788]: WARNING nova.compute.manager [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Received unexpected event network-vif-plugged-3ace6194-8918-410f-a12e-0f966ea6e08e for instance with vm_state building and task_state spawning. [ 1099.531275] env[65788]: DEBUG nova.compute.manager [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Received event network-changed-3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1099.531453] env[65788]: DEBUG nova.compute.manager [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Refreshing instance network info cache due to event network-changed-3ace6194-8918-410f-a12e-0f966ea6e08e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1099.531629] env[65788]: DEBUG oslo_concurrency.lockutils [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Acquiring lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.636201] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525258e0-bfb4-43b1-7281-22207a8f66fd, 'name': SearchDatastore_Task, 'duration_secs': 0.011895} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.636994] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-061e30e0-d3f8-4898-aace-1e923074452d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.643165] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1099.643165] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526436b4-3977-3b2f-5c06-4d9ec8da391a" [ 1099.643165] env[65788]: _type = "Task" [ 1099.643165] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.656019] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526436b4-3977-3b2f-5c06-4d9ec8da391a, 'name': SearchDatastore_Task, 'duration_secs': 0.010066} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.656301] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.656561] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1099.656842] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07ef571b-67c9-47d4-a63d-06ab3427cbec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.665373] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1099.665373] env[65788]: value = "task-4663093" [ 1099.665373] env[65788]: _type = "Task" [ 1099.665373] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.675743] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.696160] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.698860] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.397s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1099.699157] env[65788]: DEBUG nova.objects.instance [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lazy-loading 'resources' on Instance uuid 2d74b823-6e28-444d-a80b-c91c2d595460 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.714060] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1099.714461] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Instance network_info: |[{"id": "3ace6194-8918-410f-a12e-0f966ea6e08e", "address": "fa:16:3e:df:f6:3a", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ace6194-89", "ovs_interfaceid": "3ace6194-8918-410f-a12e-0f966ea6e08e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1099.715256] env[65788]: DEBUG oslo_concurrency.lockutils [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Acquired lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1099.715377] env[65788]: DEBUG nova.network.neutron [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Refreshing network info cache for port 3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1099.716587] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:f6:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '706c9762-1cf8-4770-897d-377d0d927773', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ace6194-8918-410f-a12e-0f966ea6e08e', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1099.725562] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1099.726677] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1099.726926] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c91e36fa-e1f0-4687-8834-cd54c794336d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.744378] env[65788]: INFO nova.scheduler.client.report [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleted allocations for instance c5dfbb9b-430d-4e93-b24d-e918d90e123e [ 1099.757854] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1099.757854] env[65788]: value = "task-4663094" [ 1099.757854] env[65788]: _type = "Task" [ 1099.757854] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.766090] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663094, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.823635] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1099.859658] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1099.859900] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1099.860599] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1099.860599] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1099.860599] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1099.860776] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1099.860944] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1099.861154] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1099.861322] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1099.861482] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1099.861713] env[65788]: DEBUG nova.virt.hardware [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1099.862800] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2557c3-c2fd-40f7-af12-bd3a5b92bac1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.873860] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4151aaa-3e9a-4fc0-9889-e116be79abd1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.087361] env[65788]: WARNING neutronclient.v2_0.client [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.130983] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.131214] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.131461] env[65788]: DEBUG nova.network.neutron [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1100.176831] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663093, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.227503] env[65788]: WARNING neutronclient.v2_0.client [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.228238] env[65788]: WARNING openstack [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.228588] env[65788]: WARNING openstack [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.253223] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a33ef799-643a-4229-8c8a-9e6d6647259b tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "c5dfbb9b-430d-4e93-b24d-e918d90e123e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.542s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.255687] env[65788]: DEBUG nova.network.neutron [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Successfully updated port: 01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1100.267497] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663094, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.349639] env[65788]: WARNING openstack [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.349951] env[65788]: WARNING openstack [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.418489] env[65788]: WARNING neutronclient.v2_0.client [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.419173] env[65788]: WARNING openstack [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.419524] env[65788]: WARNING openstack [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.469576] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d9d8d3-08b6-4261-a0de-7420df931a8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.478734] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865b49ad-cd43-492c-8d69-d258e7b938f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.512983] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3068e1fe-4372-49d2-aeed-cc720c44400d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.517738] env[65788]: DEBUG nova.network.neutron [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updated VIF entry in instance network info cache for port 3ace6194-8918-410f-a12e-0f966ea6e08e. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1100.518090] env[65788]: DEBUG nova.network.neutron [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updating instance_info_cache with network_info: [{"id": "3ace6194-8918-410f-a12e-0f966ea6e08e", "address": "fa:16:3e:df:f6:3a", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ace6194-89", "ovs_interfaceid": "3ace6194-8918-410f-a12e-0f966ea6e08e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1100.525145] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9702630e-4e8c-4e7b-854d-bb6ce1ba8b0f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.542850] env[65788]: DEBUG nova.compute.provider_tree [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.635079] env[65788]: WARNING neutronclient.v2_0.client [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.635480] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.635888] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.676775] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521791} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.677059] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1100.677280] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.677539] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bbcf111f-1b86-4d5e-8186-eb57f699020c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.689059] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1100.689059] env[65788]: value = "task-4663095" [ 1100.689059] env[65788]: _type = "Task" [ 1100.689059] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.699469] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.755076] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.755562] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.763064] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-068bb374-08a4-4f36-a9fc-bb4082f9eceb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.763264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-068bb374-08a4-4f36-a9fc-bb4082f9eceb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.763464] env[65788]: DEBUG nova.network.neutron [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1100.775393] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663094, 'name': CreateVM_Task, 'duration_secs': 0.674051} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.776232] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1100.776776] env[65788]: WARNING neutronclient.v2_0.client [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.777160] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.777315] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1100.777643] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1100.778272] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-827a456d-6527-43ca-b1e8-7628e912e19c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.785420] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1100.785420] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526dc898-6030-b1d8-b764-84c4ea730e83" [ 1100.785420] env[65788]: _type = "Task" [ 1100.785420] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.799095] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526dc898-6030-b1d8-b764-84c4ea730e83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.830272] env[65788]: WARNING neutronclient.v2_0.client [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1100.830951] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1100.831465] env[65788]: WARNING openstack [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1100.913612] env[65788]: DEBUG nova.network.neutron [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1101.020747] env[65788]: DEBUG oslo_concurrency.lockutils [req-77bc40ba-358c-47a1-8507-88a4e2c0280f req-26d22108-298c-4524-89b0-7e07f9f2198d service nova] Releasing lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.048999] env[65788]: DEBUG nova.scheduler.client.report [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1101.201506] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079061} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.202020] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.202971] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091c16be-3da2-45de-b0fc-b159321bf8b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.229019] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.229376] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-514cb41c-d594-4360-b451-2ddd6ea8d632 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.249855] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1101.249855] env[65788]: value = "task-4663096" [ 1101.249855] env[65788]: _type = "Task" [ 1101.249855] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.261266] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663096, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.270498] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.270966] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.297298] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526dc898-6030-b1d8-b764-84c4ea730e83, 'name': SearchDatastore_Task, 'duration_secs': 0.009774} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.297689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.297986] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1101.298322] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.298524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1101.298803] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.299168] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5ecf440-497a-4f50-9379-2eb547803671 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.312849] env[65788]: DEBUG nova.network.neutron [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1101.315386] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.315590] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1101.317870] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e0cfe4f-2052-4401-aec2-562a3312bc3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.324267] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1101.324267] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5270e2bb-d55e-304e-7d6a-454c1e613ce3" [ 1101.324267] env[65788]: _type = "Task" [ 1101.324267] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.335982] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5270e2bb-d55e-304e-7d6a-454c1e613ce3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.337613] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.338136] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.410863] env[65788]: WARNING neutronclient.v2_0.client [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1101.411642] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1101.412990] env[65788]: WARNING openstack [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1101.419756] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1101.520435] env[65788]: DEBUG nova.network.neutron [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Updating instance_info_cache with network_info: [{"id": "01a9f4ea-b2f3-4855-a441-4a5a0972a960", "address": "fa:16:3e:16:33:ed", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01a9f4ea-b2", "ovs_interfaceid": "01a9f4ea-b2f3-4855-a441-4a5a0972a960", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1101.551515] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.853s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.555515] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.909s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.555769] env[65788]: DEBUG nova.objects.instance [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lazy-loading 'resources' on Instance uuid a022cb81-5ff8-41aa-9307-c99a58416e94 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.557684] env[65788]: DEBUG nova.compute.manager [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Received event network-vif-plugged-01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1101.557880] env[65788]: DEBUG oslo_concurrency.lockutils [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Acquiring lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.558094] env[65788]: DEBUG oslo_concurrency.lockutils [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.558261] env[65788]: DEBUG oslo_concurrency.lockutils [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.559438] env[65788]: DEBUG nova.compute.manager [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] No waiting events found dispatching network-vif-plugged-01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1101.559438] env[65788]: WARNING nova.compute.manager [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Received unexpected event network-vif-plugged-01a9f4ea-b2f3-4855-a441-4a5a0972a960 for instance with vm_state building and task_state spawning. [ 1101.559438] env[65788]: DEBUG nova.compute.manager [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Received event network-changed-01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1101.559438] env[65788]: DEBUG nova.compute.manager [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Refreshing instance network info cache due to event network-changed-01a9f4ea-b2f3-4855-a441-4a5a0972a960. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1101.559438] env[65788]: DEBUG oslo_concurrency.lockutils [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Acquiring lock "refresh_cache-068bb374-08a4-4f36-a9fc-bb4082f9eceb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.629880] env[65788]: INFO nova.scheduler.client.report [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted allocations for instance 2d74b823-6e28-444d-a80b-c91c2d595460 [ 1101.761061] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663096, 'name': ReconfigVM_Task, 'duration_secs': 0.42181} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.761061] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1101.761373] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12aca009-ca9d-4a02-b496-0f069989fa56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.768381] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1101.768381] env[65788]: value = "task-4663097" [ 1101.768381] env[65788]: _type = "Task" [ 1101.768381] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.778988] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663097, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.817407] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "8755c56a-0f5e-4e95-a8b8-c643849fa798" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.817719] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.817963] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "8755c56a-0f5e-4e95-a8b8-c643849fa798-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.818229] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.818419] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.821274] env[65788]: INFO nova.compute.manager [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Terminating instance [ 1101.835991] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5270e2bb-d55e-304e-7d6a-454c1e613ce3, 'name': SearchDatastore_Task, 'duration_secs': 0.011511} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.836852] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a0506cf-f6cd-45b8-aa2e-e7c3f3412f9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.843590] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1101.843590] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be2c5f-164f-52b3-ad77-383b46954982" [ 1101.843590] env[65788]: _type = "Task" [ 1101.843590] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.853179] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be2c5f-164f-52b3-ad77-383b46954982, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.885446] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "398791b2-9dc0-478a-8ed8-bdfbace0404e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.885715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.885935] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "398791b2-9dc0-478a-8ed8-bdfbace0404e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1101.886139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.886313] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.888458] env[65788]: INFO nova.compute.manager [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Terminating instance [ 1101.929771] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e0a045-a4e5-4ff9-9cc1-31b008e61f2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.938342] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84b5464-655e-418f-a6c9-b1f49a46c8fb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.023985] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-068bb374-08a4-4f36-a9fc-bb4082f9eceb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.024972] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Instance network_info: |[{"id": "01a9f4ea-b2f3-4855-a441-4a5a0972a960", "address": "fa:16:3e:16:33:ed", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01a9f4ea-b2", "ovs_interfaceid": "01a9f4ea-b2f3-4855-a441-4a5a0972a960", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1102.025141] env[65788]: DEBUG oslo_concurrency.lockutils [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Acquired lock "refresh_cache-068bb374-08a4-4f36-a9fc-bb4082f9eceb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.025290] env[65788]: DEBUG nova.network.neutron [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Refreshing network info cache for port 01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1102.026575] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:33:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01a9f4ea-b2f3-4855-a441-4a5a0972a960', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1102.036043] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1102.036982] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1102.037233] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66941d7b-5b8b-46d8-85b7-e13eeb077839 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.058560] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1102.058560] env[65788]: value = "task-4663098" [ 1102.058560] env[65788]: _type = "Task" [ 1102.058560] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.069813] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663098, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.143204] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d76123-fd70-4b6d-a7ca-3329d3a070d4 tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "2d74b823-6e28-444d-a80b-c91c2d595460" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.255s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.279372] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663097, 'name': Rename_Task, 'duration_secs': 0.154739} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.282204] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1102.283416] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cf22208-5474-431d-bd8f-6788ed5b317d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.290782] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1102.290782] env[65788]: value = "task-4663099" [ 1102.290782] env[65788]: _type = "Task" [ 1102.290782] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.304029] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.309233] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af47e06-9eea-4e80-9ef3-4e652103da75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.318431] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567ac042-15e9-4002-b961-952d45014ee4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.325209] env[65788]: DEBUG nova.compute.manager [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1102.325518] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1102.357535] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772c7427-fac6-4005-be44-bbebbbd79fa9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.366657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1493294d-84e8-4778-a5d4-90d418359985 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.376927] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52be2c5f-164f-52b3-ad77-383b46954982, 'name': SearchDatastore_Task, 'duration_secs': 0.010651} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.381074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1102.381359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 9946cef4-b8c7-41b4-9299-eace1929bf3d/9946cef4-b8c7-41b4-9299-eace1929bf3d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1102.381751] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.381986] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9f8c0da-05d7-4bb8-82da-306f704b83ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.385118] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5384b74c-3218-4b88-89c6-e355fb66dbc8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.389497] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bc379fa-c20e-4a33-bc95-40fc5975d8f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.392282] env[65788]: DEBUG nova.compute.manager [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1102.392548] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1102.393349] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1eb717-2186-4c10-8334-f3d6f8b764a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.408281] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1102.408281] env[65788]: value = "task-4663101" [ 1102.408281] env[65788]: _type = "Task" [ 1102.408281] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.408573] env[65788]: DEBUG nova.compute.provider_tree [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.410396] env[65788]: DEBUG oslo_vmware.api [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1102.410396] env[65788]: value = "task-4663100" [ 1102.410396] env[65788]: _type = "Task" [ 1102.410396] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.423743] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.423743] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47c04f35-497d-4b4a-b11d-10ada5e62493 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.432751] env[65788]: DEBUG oslo_vmware.api [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663100, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.433521] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663101, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.435025] env[65788]: DEBUG oslo_vmware.api [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1102.435025] env[65788]: value = "task-4663102" [ 1102.435025] env[65788]: _type = "Task" [ 1102.435025] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.445689] env[65788]: DEBUG oslo_vmware.api [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663102, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.537451] env[65788]: WARNING neutronclient.v2_0.client [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.539284] env[65788]: WARNING openstack [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.540122] env[65788]: WARNING openstack [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.571737] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663098, 'name': CreateVM_Task, 'duration_secs': 0.367394} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.572191] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1102.572854] env[65788]: WARNING neutronclient.v2_0.client [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.573356] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.573574] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1102.574114] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1102.574773] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f37d3f2c-3c8a-4834-aed7-3ce4cda3ab5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.582937] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1102.582937] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d31-cfcf-83a2-3f5e-8b793f6f197c" [ 1102.582937] env[65788]: _type = "Task" [ 1102.582937] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.595796] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d31-cfcf-83a2-3f5e-8b793f6f197c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.673747] env[65788]: WARNING openstack [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.674216] env[65788]: WARNING openstack [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.747446] env[65788]: WARNING neutronclient.v2_0.client [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1102.748183] env[65788]: WARNING openstack [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1102.748536] env[65788]: WARNING openstack [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1102.809782] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663099, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.839636] env[65788]: DEBUG nova.network.neutron [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Updated VIF entry in instance network info cache for port 01a9f4ea-b2f3-4855-a441-4a5a0972a960. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1102.840029] env[65788]: DEBUG nova.network.neutron [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Updating instance_info_cache with network_info: [{"id": "01a9f4ea-b2f3-4855-a441-4a5a0972a960", "address": "fa:16:3e:16:33:ed", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01a9f4ea-b2", "ovs_interfaceid": "01a9f4ea-b2f3-4855-a441-4a5a0972a960", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1102.912515] env[65788]: DEBUG nova.scheduler.client.report [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1102.932558] env[65788]: DEBUG oslo_vmware.api [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663100, 'name': PowerOffVM_Task, 'duration_secs': 0.182943} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.935548] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.935725] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1102.936281] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663101, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520694} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.936490] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8416c4ff-d584-48e6-8dd6-d1eab400e2a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.940859] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 9946cef4-b8c7-41b4-9299-eace1929bf3d/9946cef4-b8c7-41b4-9299-eace1929bf3d.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1102.941077] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1102.941791] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97deffa9-6771-4a17-a181-e8cf4a0cd5a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.949109] env[65788]: DEBUG oslo_vmware.api [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663102, 'name': PowerOffVM_Task, 'duration_secs': 0.204892} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.949363] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1102.949534] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1102.950257] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bfa841d-13d9-4705-b06a-a66f6b1a192d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.952799] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1102.952799] env[65788]: value = "task-4663104" [ 1102.952799] env[65788]: _type = "Task" [ 1102.952799] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.964290] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663104, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.056583] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1103.056950] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1103.057319] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleting the datastore file [datastore2] 8755c56a-0f5e-4e95-a8b8-c643849fa798 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.058481] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bc1ee3b0-21e9-45a8-9f2e-f28db405d137 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.062429] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1103.062756] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1103.062996] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleting the datastore file [datastore1] 398791b2-9dc0-478a-8ed8-bdfbace0404e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.063921] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-081fdb3e-23f5-4d8b-b408-d4d34510df9b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.069130] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b5498d-5dbf-481e-b3bc-ac696e3c2f19 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.074887] env[65788]: DEBUG oslo_vmware.api [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1103.074887] env[65788]: value = "task-4663106" [ 1103.074887] env[65788]: _type = "Task" [ 1103.074887] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.097729] env[65788]: DEBUG oslo_vmware.api [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for the task: (returnval){ [ 1103.097729] env[65788]: value = "task-4663107" [ 1103.097729] env[65788]: _type = "Task" [ 1103.097729] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.102458] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4c7935-b89a-48ef-9777-0147756c29c1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.112877] env[65788]: DEBUG oslo_vmware.api [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.120641] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52620d31-cfcf-83a2-3f5e-8b793f6f197c, 'name': SearchDatastore_Task, 'duration_secs': 0.071993} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.126766] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.127063] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1103.127332] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.127488] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.127696] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1103.127999] env[65788]: DEBUG oslo_vmware.api [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663107, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.128451] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1103.132440] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc8e018f-7a51-4661-a45c-ee3cf009af26 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.142900] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1103.143140] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1103.143976] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b49a70f2-9600-4dcb-b9f8-1576d5b83dd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.150193] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1103.150193] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f4118-a9a5-8eb9-8d54-5b815c275509" [ 1103.150193] env[65788]: _type = "Task" [ 1103.150193] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.158644] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f4118-a9a5-8eb9-8d54-5b815c275509, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.303085] env[65788]: DEBUG oslo_vmware.api [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663099, 'name': PowerOnVM_Task, 'duration_secs': 0.826594} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.303543] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1103.303819] env[65788]: INFO nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Took 8.26 seconds to spawn the instance on the hypervisor. [ 1103.304031] env[65788]: DEBUG nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1103.304961] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b622eca-46c7-4b2b-8b20-a09bacb2af66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.342953] env[65788]: DEBUG oslo_concurrency.lockutils [req-94da1b24-3211-4f6a-885a-42b805a7ed8f req-3d89a1b0-ace1-4724-876b-f3b6285b8c11 service nova] Releasing lock "refresh_cache-068bb374-08a4-4f36-a9fc-bb4082f9eceb" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1103.427443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.872s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.456224] env[65788]: INFO nova.scheduler.client.report [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Deleted allocations for instance a022cb81-5ff8-41aa-9307-c99a58416e94 [ 1103.472179] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663104, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073201} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.472376] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1103.473261] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e658baae-5d6a-4087-9769-1fbc7ba28b4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.499770] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 9946cef4-b8c7-41b4-9299-eace1929bf3d/9946cef4-b8c7-41b4-9299-eace1929bf3d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1103.500293] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96d6a070-c981-4896-9d72-c5f590c3577d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.523971] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1103.523971] env[65788]: value = "task-4663108" [ 1103.523971] env[65788]: _type = "Task" [ 1103.523971] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.533559] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663108, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.587340] env[65788]: DEBUG oslo_vmware.api [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.427341} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.587814] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.588053] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1103.588257] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1103.588438] env[65788]: INFO nova.compute.manager [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Took 1.26 seconds to destroy the instance on the hypervisor. [ 1103.588718] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1103.588940] env[65788]: DEBUG nova.compute.manager [-] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1103.589058] env[65788]: DEBUG nova.network.neutron [-] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1103.589332] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.589892] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.590168] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.617175] env[65788]: DEBUG oslo_vmware.api [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Task: {'id': task-4663107, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413674} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.617588] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1103.617852] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1103.618119] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1103.618335] env[65788]: INFO nova.compute.manager [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1103.618612] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1103.618835] env[65788]: DEBUG nova.compute.manager [-] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1103.618935] env[65788]: DEBUG nova.network.neutron [-] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1103.619214] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.619853] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1103.620127] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1103.632263] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.637997] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1103.638328] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bfd26f3d-fd98-4f24-8f9e-2948d666f728 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.647154] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1103.647154] env[65788]: value = "task-4663109" [ 1103.647154] env[65788]: _type = "Task" [ 1103.647154] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.660990] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.665335] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f4118-a9a5-8eb9-8d54-5b815c275509, 'name': SearchDatastore_Task, 'duration_secs': 0.021101} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.666276] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c2c5c0b-6142-4cdc-b7af-d7e29072cf76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.670504] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1103.674031] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1103.674031] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521287db-5458-a475-c93e-6db9382b4985" [ 1103.674031] env[65788]: _type = "Task" [ 1103.674031] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.683696] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521287db-5458-a475-c93e-6db9382b4985, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.827043] env[65788]: INFO nova.compute.manager [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Took 36.17 seconds to build instance. [ 1103.898980] env[65788]: DEBUG nova.compute.manager [req-e611b427-6a76-4580-9502-0dd2b888325d req-5edd4f9e-f17d-4d91-9dac-f14deed11552 service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Received event network-vif-deleted-c26e3ef9-b46e-47db-8ad1-682fc4c31a62 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1103.899203] env[65788]: INFO nova.compute.manager [req-e611b427-6a76-4580-9502-0dd2b888325d req-5edd4f9e-f17d-4d91-9dac-f14deed11552 service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Neutron deleted interface c26e3ef9-b46e-47db-8ad1-682fc4c31a62; detaching it from the instance and deleting it from the info cache [ 1103.899538] env[65788]: DEBUG nova.network.neutron [req-e611b427-6a76-4580-9502-0dd2b888325d req-5edd4f9e-f17d-4d91-9dac-f14deed11552 service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.918861] env[65788]: DEBUG nova.compute.manager [req-45342bf2-1d46-4712-bf76-9078af5176f8 req-94018e37-bd05-4a66-b800-7171d6cdb212 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Received event network-vif-deleted-ef95980e-7eaa-4733-ae0d-445fe80490ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1103.918861] env[65788]: INFO nova.compute.manager [req-45342bf2-1d46-4712-bf76-9078af5176f8 req-94018e37-bd05-4a66-b800-7171d6cdb212 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Neutron deleted interface ef95980e-7eaa-4733-ae0d-445fe80490ea; detaching it from the instance and deleting it from the info cache [ 1103.918861] env[65788]: DEBUG nova.network.neutron [req-45342bf2-1d46-4712-bf76-9078af5176f8 req-94018e37-bd05-4a66-b800-7171d6cdb212 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1103.969013] env[65788]: DEBUG oslo_concurrency.lockutils [None req-00dede9a-4834-4a0b-9086-1e6cd28d0e8e tempest-MultipleCreateTestJSON-1108671505 tempest-MultipleCreateTestJSON-1108671505-project-member] Lock "a022cb81-5ff8-41aa-9307-c99a58416e94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.025s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.035040] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663108, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.159170] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663109, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.185992] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521287db-5458-a475-c93e-6db9382b4985, 'name': SearchDatastore_Task, 'duration_secs': 0.020918} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.186476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.187967] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 068bb374-08a4-4f36-a9fc-bb4082f9eceb/068bb374-08a4-4f36-a9fc-bb4082f9eceb.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1104.187967] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ad889b7-dcb7-4913-906c-6e96ddd093ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.197268] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1104.197268] env[65788]: value = "task-4663110" [ 1104.197268] env[65788]: _type = "Task" [ 1104.197268] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.207030] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663110, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.327430] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d9b8e23e-8526-4f02-b9d8-a2c0316d2b2c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.684s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.347083] env[65788]: DEBUG nova.network.neutron [-] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.355397] env[65788]: DEBUG nova.network.neutron [-] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1104.408049] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71a26493-ece0-4f46-9833-983c0efdd1d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.418126] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e97a41-baa7-4b62-b28d-faf11a66c0f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.429576] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c73de9d-027c-42ed-b495-f75c95cfbfb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.439985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9fc2fc-4ad2-4650-baa4-381087ee4dc9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.463612] env[65788]: DEBUG nova.compute.manager [req-e611b427-6a76-4580-9502-0dd2b888325d req-5edd4f9e-f17d-4d91-9dac-f14deed11552 service nova] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Detach interface failed, port_id=c26e3ef9-b46e-47db-8ad1-682fc4c31a62, reason: Instance 8755c56a-0f5e-4e95-a8b8-c643849fa798 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1104.493093] env[65788]: DEBUG nova.compute.manager [req-45342bf2-1d46-4712-bf76-9078af5176f8 req-94018e37-bd05-4a66-b800-7171d6cdb212 service nova] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Detach interface failed, port_id=ef95980e-7eaa-4733-ae0d-445fe80490ea, reason: Instance 398791b2-9dc0-478a-8ed8-bdfbace0404e could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1104.535603] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663108, 'name': ReconfigVM_Task, 'duration_secs': 0.81157} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.535913] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 9946cef4-b8c7-41b4-9299-eace1929bf3d/9946cef4-b8c7-41b4-9299-eace1929bf3d.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1104.536603] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c080e636-b9a1-477c-adef-28e9054c23d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.544878] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1104.544878] env[65788]: value = "task-4663111" [ 1104.544878] env[65788]: _type = "Task" [ 1104.544878] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.553731] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663111, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.659252] env[65788]: DEBUG oslo_vmware.api [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663109, 'name': PowerOnVM_Task, 'duration_secs': 0.695977} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.659252] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1104.659252] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2264cf93-944a-4b7a-b9e0-05f986a957af tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance '18be7322-3359-49c9-a181-f9228d5c16d7' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1104.709665] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663110, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.850739] env[65788]: INFO nova.compute.manager [-] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Took 1.26 seconds to deallocate network for instance. [ 1104.859552] env[65788]: INFO nova.compute.manager [-] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Took 1.24 seconds to deallocate network for instance. [ 1104.872512] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.872788] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.872996] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.873194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.873371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1104.877351] env[65788]: INFO nova.compute.manager [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Terminating instance [ 1105.059074] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663111, 'name': Rename_Task, 'duration_secs': 0.503734} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.059654] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1105.060137] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fc2cf93-655f-42c7-9c5b-cab0f08a4ccd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.069982] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1105.069982] env[65788]: value = "task-4663112" [ 1105.069982] env[65788]: _type = "Task" [ 1105.069982] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.078167] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663112, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.210726] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663110, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.899168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.212808] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 068bb374-08a4-4f36-a9fc-bb4082f9eceb/068bb374-08a4-4f36-a9fc-bb4082f9eceb.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1105.212808] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1105.212808] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19696233-2a8c-4951-9747-58679d4684f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.222666] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1105.222666] env[65788]: value = "task-4663113" [ 1105.222666] env[65788]: _type = "Task" [ 1105.222666] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.233045] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.357445] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.357632] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.358480] env[65788]: DEBUG nova.objects.instance [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lazy-loading 'resources' on Instance uuid 8755c56a-0f5e-4e95-a8b8-c643849fa798 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1105.370685] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.386322] env[65788]: DEBUG nova.compute.manager [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1105.386322] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1105.391174] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75dba35-3b28-489d-a9a7-84cc452bf444 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.400168] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.400168] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cd85b51-e98c-4253-a5a1-c99143dcfc53 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.410301] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1105.410301] env[65788]: value = "task-4663114" [ 1105.410301] env[65788]: _type = "Task" [ 1105.410301] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.422790] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.587045] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663112, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.735601] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.184503} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.735960] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1105.736772] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e9905c-de13-496f-b5ff-8aa33258ef96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.765216] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 068bb374-08a4-4f36-a9fc-bb4082f9eceb/068bb374-08a4-4f36-a9fc-bb4082f9eceb.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1105.765592] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-feb71d3c-386d-4edc-9f78-3a241ccf9d4f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.790396] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1105.790396] env[65788]: value = "task-4663115" [ 1105.790396] env[65788]: _type = "Task" [ 1105.790396] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.799754] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663115, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.925549] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663114, 'name': PowerOffVM_Task, 'duration_secs': 0.207201} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.925840] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.926070] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1105.926337] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-06377bb9-9c3d-43bc-a194-90c5dddebee0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.943050] env[65788]: DEBUG nova.compute.manager [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1105.943181] env[65788]: DEBUG nova.compute.manager [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing instance network info cache due to event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1105.945321] env[65788]: DEBUG oslo_concurrency.lockutils [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.945321] env[65788]: DEBUG oslo_concurrency.lockutils [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1105.945321] env[65788]: DEBUG nova.network.neutron [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1106.013801] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1106.014461] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1106.014664] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleting the datastore file [datastore2] bc39a402-1f5a-4a67-b09f-a18159cf4abf {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1106.015017] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ab71a6a-f563-4e04-9c46-93b09ccf736a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.028828] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1106.028828] env[65788]: value = "task-4663117" [ 1106.028828] env[65788]: _type = "Task" [ 1106.028828] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.041358] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663117, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.081821] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663112, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.134768] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5360df-74ef-4433-b2dc-7ba020a90cd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.144972] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbaaed7-98dd-4013-9477-5eba54e10c4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.182986] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d9e60c-609b-4cde-ade7-e5cdd0879992 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.192481] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f1cf44-6832-41fd-af9f-d062dec6db29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.208534] env[65788]: DEBUG nova.compute.provider_tree [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1106.302624] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663115, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.447622] env[65788]: WARNING neutronclient.v2_0.client [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1106.448578] env[65788]: WARNING openstack [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.449307] env[65788]: WARNING openstack [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.540534] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663117, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.580344] env[65788]: DEBUG oslo_vmware.api [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663112, 'name': PowerOnVM_Task, 'duration_secs': 1.075527} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.580894] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1106.580894] env[65788]: INFO nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Took 9.16 seconds to spawn the instance on the hypervisor. [ 1106.581441] env[65788]: DEBUG nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1106.581978] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4603e1-a711-49ca-80bb-67b4d09e697b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.599863] env[65788]: WARNING openstack [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.600493] env[65788]: WARNING openstack [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.668210] env[65788]: WARNING neutronclient.v2_0.client [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1106.668944] env[65788]: WARNING openstack [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1106.669538] env[65788]: WARNING openstack [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1106.747301] env[65788]: DEBUG nova.scheduler.client.report [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 141 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1106.747729] env[65788]: DEBUG nova.compute.provider_tree [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 141 to 142 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1106.748037] env[65788]: DEBUG nova.compute.provider_tree [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1106.790281] env[65788]: DEBUG nova.network.neutron [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updated VIF entry in instance network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1106.790666] env[65788]: DEBUG nova.network.neutron [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1106.805314] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663115, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.902521] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "18be7322-3359-49c9-a181-f9228d5c16d7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.902808] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.903140] env[65788]: DEBUG nova.compute.manager [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Going to confirm migration 7 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1107.043123] env[65788]: DEBUG oslo_vmware.api [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663117, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.871451} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.043848] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.044296] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.045015] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.045200] env[65788]: INFO nova.compute.manager [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Took 1.66 seconds to destroy the instance on the hypervisor. [ 1107.045776] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1107.045983] env[65788]: DEBUG nova.compute.manager [-] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1107.046216] env[65788]: DEBUG nova.network.neutron [-] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1107.046897] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.047864] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1107.048328] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1107.096263] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.106279] env[65788]: INFO nova.compute.manager [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Took 22.73 seconds to build instance. [ 1107.255415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.898s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.258415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.888s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1107.258660] env[65788]: DEBUG nova.objects.instance [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lazy-loading 'resources' on Instance uuid 398791b2-9dc0-478a-8ed8-bdfbace0404e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.285940] env[65788]: INFO nova.scheduler.client.report [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleted allocations for instance 8755c56a-0f5e-4e95-a8b8-c643849fa798 [ 1107.299099] env[65788]: DEBUG oslo_concurrency.lockutils [req-46200fb7-5831-4e7c-b611-a3c91de58604 req-15ef983e-5141-4b7b-b963-8b65789c3a02 service nova] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1107.307861] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663115, 'name': ReconfigVM_Task, 'duration_secs': 1.404325} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.308596] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 068bb374-08a4-4f36-a9fc-bb4082f9eceb/068bb374-08a4-4f36-a9fc-bb4082f9eceb.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.309141] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d196cdfd-928b-4996-a72a-96f5821c2007 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.318062] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1107.318062] env[65788]: value = "task-4663118" [ 1107.318062] env[65788]: _type = "Task" [ 1107.318062] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.327235] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663118, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.410235] env[65788]: WARNING neutronclient.v2_0.client [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.608593] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b766ddfb-6e7f-48f0-b6ee-a7303fa0fa00 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.247s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.614275] env[65788]: WARNING neutronclient.v2_0.client [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1107.614632] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.614786] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquired lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1107.614978] env[65788]: DEBUG nova.network.neutron [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1107.615250] env[65788]: DEBUG nova.objects.instance [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'info_cache' on Instance uuid 18be7322-3359-49c9-a181-f9228d5c16d7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.734767] env[65788]: DEBUG nova.compute.manager [req-cafc3502-d00d-4407-ac9a-016047ff99dc req-8097e55a-0a99-4fbb-9d21-e6acbc73ffd8 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Received event network-vif-deleted-d6d70bc9-e4ba-4775-9452-626cf247086d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1107.735019] env[65788]: INFO nova.compute.manager [req-cafc3502-d00d-4407-ac9a-016047ff99dc req-8097e55a-0a99-4fbb-9d21-e6acbc73ffd8 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Neutron deleted interface d6d70bc9-e4ba-4775-9452-626cf247086d; detaching it from the instance and deleting it from the info cache [ 1107.736519] env[65788]: DEBUG nova.network.neutron [req-cafc3502-d00d-4407-ac9a-016047ff99dc req-8097e55a-0a99-4fbb-9d21-e6acbc73ffd8 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1107.794990] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05bd3db-d1e9-4156-b0db-e3e716d7ecb1 tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "8755c56a-0f5e-4e95-a8b8-c643849fa798" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.977s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1107.830174] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663118, 'name': Rename_Task, 'duration_secs': 0.277412} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.830823] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1107.831098] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91add674-527c-4536-9b1f-3115be3c8e5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.842770] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1107.842770] env[65788]: value = "task-4663119" [ 1107.842770] env[65788]: _type = "Task" [ 1107.842770] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.854267] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663119, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.994248] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5b98cd-027d-4da3-951a-1fe69599ada5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.004104] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fda162-ef8d-4046-85d8-3027b1f6a9c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.045382] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987eca8c-b3cc-48f5-bb3d-fcd4e350de81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.056119] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ef5157-d1e3-45c4-a174-a598fbe842a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.071490] env[65788]: DEBUG nova.compute.provider_tree [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.072521] env[65788]: DEBUG nova.network.neutron [-] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1108.238486] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5488f1b1-4f7f-4ded-8a16-10e97c494d9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.252147] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73610ca4-ac33-41df-b524-472ab7857e40 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.296960] env[65788]: DEBUG nova.compute.manager [req-cafc3502-d00d-4407-ac9a-016047ff99dc req-8097e55a-0a99-4fbb-9d21-e6acbc73ffd8 service nova] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Detach interface failed, port_id=d6d70bc9-e4ba-4775-9452-626cf247086d, reason: Instance bc39a402-1f5a-4a67-b09f-a18159cf4abf could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1108.354085] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663119, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.576264] env[65788]: DEBUG nova.scheduler.client.report [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1108.579471] env[65788]: INFO nova.compute.manager [-] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Took 1.53 seconds to deallocate network for instance. [ 1108.624032] env[65788]: WARNING neutronclient.v2_0.client [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.624194] env[65788]: WARNING openstack [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.624489] env[65788]: WARNING openstack [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.805980] env[65788]: WARNING openstack [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.806484] env[65788]: WARNING openstack [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.858772] env[65788]: DEBUG oslo_vmware.api [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663119, 'name': PowerOnVM_Task, 'duration_secs': 0.940626} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.858772] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.858772] env[65788]: INFO nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Took 9.03 seconds to spawn the instance on the hypervisor. [ 1108.858772] env[65788]: DEBUG nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1108.859713] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a58c600-6200-4835-b8cb-2d83601f473f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.882877] env[65788]: WARNING neutronclient.v2_0.client [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1108.883630] env[65788]: WARNING openstack [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1108.884298] env[65788]: WARNING openstack [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1108.975206] env[65788]: DEBUG nova.network.neutron [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [{"id": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "address": "fa:16:3e:c4:f5:65", "network": {"id": "a76ab1ac-3f2a-4812-9181-bd7bc0fb625a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1983874800-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ffbe44d63c1d432e97849f15615329e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "099fe970-c61f-4480-bed4-ae4f485fd82a", "external-id": "nsx-vlan-transportzone-678", "segmentation_id": 678, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab03f5bb-a9", "ovs_interfaceid": "ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1109.081236] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.088152] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1109.088466] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1109.088687] env[65788]: DEBUG nova.objects.instance [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'resources' on Instance uuid bc39a402-1f5a-4a67-b09f-a18159cf4abf {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.107296] env[65788]: INFO nova.scheduler.client.report [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Deleted allocations for instance 398791b2-9dc0-478a-8ed8-bdfbace0404e [ 1109.381244] env[65788]: INFO nova.compute.manager [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Took 21.16 seconds to build instance. [ 1109.478228] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Releasing lock "refresh_cache-18be7322-3359-49c9-a181-f9228d5c16d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.478228] env[65788]: DEBUG nova.objects.instance [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'migration_context' on Instance uuid 18be7322-3359-49c9-a181-f9228d5c16d7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.617315] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d969429a-d0a0-49a8-b8ff-4357f78577eb tempest-ListServersNegativeTestJSON-1635349101 tempest-ListServersNegativeTestJSON-1635349101-project-member] Lock "398791b2-9dc0-478a-8ed8-bdfbace0404e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.730s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.764614] env[65788]: DEBUG nova.compute.manager [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Received event network-changed-3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1109.764801] env[65788]: DEBUG nova.compute.manager [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Refreshing instance network info cache due to event network-changed-3ace6194-8918-410f-a12e-0f966ea6e08e. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1109.765035] env[65788]: DEBUG oslo_concurrency.lockutils [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Acquiring lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.765676] env[65788]: DEBUG oslo_concurrency.lockutils [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Acquired lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.765676] env[65788]: DEBUG nova.network.neutron [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Refreshing network info cache for port 3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1109.803692] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59216f3-4163-4d5a-bdcb-0a440eb71d83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.813356] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d508c9-c68e-4e1f-a9ad-1d9ed113307b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.847307] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfc8f71-1c06-4c30-8d77-ff4375379c76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.856032] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7077c690-1a4f-4eba-adb9-42fcd30c58ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.871439] env[65788]: DEBUG nova.compute.provider_tree [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.882958] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbb015a7-f1f5-4d76-8200-bf75668ae666 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.669s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1109.982310] env[65788]: DEBUG nova.objects.base [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Object Instance<18be7322-3359-49c9-a181-f9228d5c16d7> lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1109.984428] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51b7cdb-56ee-4b81-b4af-8a9eae861164 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.007248] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62dd3750-47c5-42f6-985b-8f7d9cf035e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.014262] env[65788]: DEBUG oslo_vmware.api [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1110.014262] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271fa68-c230-427c-5a7b-ee35308edfa8" [ 1110.014262] env[65788]: _type = "Task" [ 1110.014262] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.023791] env[65788]: DEBUG oslo_vmware.api [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271fa68-c230-427c-5a7b-ee35308edfa8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.163563] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1110.164457] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.171888] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b30027-18cf-42f1-9f33-bab82835a70e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.180515] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Suspending the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1110.181430] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-b2daed21-94c5-4ca6-a76a-5a733c778cb6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.190046] env[65788]: DEBUG oslo_vmware.api [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1110.190046] env[65788]: value = "task-4663120" [ 1110.190046] env[65788]: _type = "Task" [ 1110.190046] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.200064] env[65788]: DEBUG oslo_vmware.api [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663120, 'name': SuspendVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.271313] env[65788]: WARNING neutronclient.v2_0.client [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.272427] env[65788]: WARNING openstack [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.273593] env[65788]: WARNING openstack [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.373346] env[65788]: DEBUG nova.scheduler.client.report [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1110.404896] env[65788]: WARNING openstack [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.405323] env[65788]: WARNING openstack [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.473242] env[65788]: WARNING neutronclient.v2_0.client [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1110.473942] env[65788]: WARNING openstack [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1110.474321] env[65788]: WARNING openstack [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1110.528894] env[65788]: DEBUG oslo_vmware.api [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5271fa68-c230-427c-5a7b-ee35308edfa8, 'name': SearchDatastore_Task, 'duration_secs': 0.014844} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.529320] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1110.572093] env[65788]: DEBUG nova.network.neutron [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updated VIF entry in instance network info cache for port 3ace6194-8918-410f-a12e-0f966ea6e08e. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1110.572520] env[65788]: DEBUG nova.network.neutron [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updating instance_info_cache with network_info: [{"id": "3ace6194-8918-410f-a12e-0f966ea6e08e", "address": "fa:16:3e:df:f6:3a", "network": {"id": "7662d96f-a4fd-47ab-bdf8-19f420a3c453", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-749789497-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "021bf6dd87e340969a57e7e5b33268e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "706c9762-1cf8-4770-897d-377d0d927773", "external-id": "nsx-vlan-transportzone-402", "segmentation_id": 402, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ace6194-89", "ovs_interfaceid": "3ace6194-8918-410f-a12e-0f966ea6e08e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1110.666758] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1110.703637] env[65788]: DEBUG oslo_vmware.api [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663120, 'name': SuspendVM_Task} progress is 58%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.879576] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.882124] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.353s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.907508] env[65788]: INFO nova.scheduler.client.report [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted allocations for instance bc39a402-1f5a-4a67-b09f-a18159cf4abf [ 1111.077452] env[65788]: DEBUG oslo_concurrency.lockutils [req-e1ee378a-4028-4af1-ad7c-417e3e927c59 req-d74edcd4-0476-4baa-a676-c09c3bba25b8 service nova] Releasing lock "refresh_cache-9946cef4-b8c7-41b4-9299-eace1929bf3d" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.199260] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.209374] env[65788]: DEBUG oslo_vmware.api [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663120, 'name': SuspendVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.419758] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9af37f8d-d49d-4bd6-84b4-d7dd64240ef2 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "bc39a402-1f5a-4a67-b09f-a18159cf4abf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.547s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.578387] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3792d44-605d-464d-aa41-6ba4dfbf3a6a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.588800] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66984163-c1ba-4afb-8aaf-c5af6fc59bb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.625082] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e14b019-e21b-4e73-b8c2-62e90c67135f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.635543] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e2840f-6d66-4d4d-93b3-684c94a5e0a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.652818] env[65788]: DEBUG nova.compute.provider_tree [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.701992] env[65788]: DEBUG oslo_vmware.api [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663120, 'name': SuspendVM_Task, 'duration_secs': 1.050379} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.702300] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Suspended the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1111.702490] env[65788]: DEBUG nova.compute.manager [None req-4ddbee9b-910b-44a0-a450-77408654818f tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1111.703422] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c0f5b7-c2db-44ee-826e-8f34c2321ae3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.157470] env[65788]: DEBUG nova.scheduler.client.report [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1112.170574] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.170756] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.629256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.629544] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.629762] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.629940] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.630128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.632592] env[65788]: INFO nova.compute.manager [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Terminating instance [ 1112.675602] env[65788]: INFO nova.compute.manager [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Detaching volume 676fc2ff-ea9f-4bd0-bd6a-86d399263acc [ 1112.718546] env[65788]: INFO nova.virt.block_device [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Attempting to driver detach volume 676fc2ff-ea9f-4bd0-bd6a-86d399263acc from mountpoint /dev/sdb [ 1112.718781] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1112.718969] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910384', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'name': 'volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ec7f7a46-8dd2-4b75-866f-20e73907f1cd', 'attached_at': '', 'detached_at': '', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'serial': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1112.720088] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bec918-7eb6-437f-a071-cd15f4774a0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.745320] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26478c8-f054-43ed-b75d-086d427d2ae8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.753782] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e088f6e0-89ac-43a7-8059-2879b207975b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.775273] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ab5029-48e7-47dd-9f4f-63ece69c9df3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.792390] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The volume has not been displaced from its original location: [datastore2] volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc/volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1112.797865] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1112.798283] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51bbc78f-87d7-402a-836e-fb3c8e63479b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.817582] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1112.817582] env[65788]: value = "task-4663121" [ 1112.817582] env[65788]: _type = "Task" [ 1112.817582] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.826787] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663121, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.137395] env[65788]: DEBUG nova.compute.manager [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1113.137722] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1113.138753] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8033f8ca-20f2-4910-a6e2-7fcf8562257c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.148770] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1113.149213] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46d972bc-ea96-4f19-9623-ee5d20553097 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.167933] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.286s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1113.171428] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.972s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1113.173170] env[65788]: INFO nova.compute.claims [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1113.229465] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1113.229673] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1113.229854] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleting the datastore file [datastore2] 068bb374-08a4-4f36-a9fc-bb4082f9eceb {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1113.230152] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a960e0ce-cfbb-4432-950a-51105612a7fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.238068] env[65788]: DEBUG oslo_vmware.api [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1113.238068] env[65788]: value = "task-4663123" [ 1113.238068] env[65788]: _type = "Task" [ 1113.238068] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.247416] env[65788]: DEBUG oslo_vmware.api [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.329889] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663121, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.557715] env[65788]: INFO nova.compute.manager [None req-09219cf0-25b9-4a8e-a440-074d06218e51 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Get console output [ 1113.558181] env[65788]: WARNING nova.virt.vmwareapi.driver [None req-09219cf0-25b9-4a8e-a440-074d06218e51 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] The console log is missing. Check your VSPC configuration [ 1113.738682] env[65788]: INFO nova.scheduler.client.report [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted allocation for migration b2330f68-7793-465a-8fed-06f3654996c8 [ 1113.752576] env[65788]: DEBUG oslo_vmware.api [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.428128} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.752830] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.755661] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1113.755661] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1113.755661] env[65788]: INFO nova.compute.manager [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1113.755661] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1113.755661] env[65788]: DEBUG nova.compute.manager [-] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1113.755661] env[65788]: DEBUG nova.network.neutron [-] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1113.755943] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.756349] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1113.757250] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1113.797564] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1113.829026] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663121, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.087330] env[65788]: DEBUG nova.compute.manager [req-8b972383-b0bf-40c0-b44f-508cb70e8bc5 req-2a011184-7c2c-461a-89b7-bba36772f7d5 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Received event network-vif-deleted-01a9f4ea-b2f3-4855-a441-4a5a0972a960 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1114.087330] env[65788]: INFO nova.compute.manager [req-8b972383-b0bf-40c0-b44f-508cb70e8bc5 req-2a011184-7c2c-461a-89b7-bba36772f7d5 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Neutron deleted interface 01a9f4ea-b2f3-4855-a441-4a5a0972a960; detaching it from the instance and deleting it from the info cache [ 1114.087560] env[65788]: DEBUG nova.network.neutron [req-8b972383-b0bf-40c0-b44f-508cb70e8bc5 req-2a011184-7c2c-461a-89b7-bba36772f7d5 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.249270] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b53c3f5d-a662-4860-888b-cc3764547b08 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.346s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.332710] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663121, 'name': ReconfigVM_Task, 'duration_secs': 1.413452} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.332941] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1114.340889] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1831d883-eee3-4a0c-b2fb-131ddc8a02fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.360680] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1114.360680] env[65788]: value = "task-4663124" [ 1114.360680] env[65788]: _type = "Task" [ 1114.360680] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.374576] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663124, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.415602] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0d0f1b-77dc-4bff-8229-1ebbb152b45e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.424575] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703a33e1-2205-4268-bf1e-043465fd9212 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.463976] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d50642-3a10-43a2-8853-ce6449127051 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.473750] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e278bc-7b4e-44ec-980b-d606d01270e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.492067] env[65788]: DEBUG nova.compute.provider_tree [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.515135] env[65788]: DEBUG nova.network.neutron [-] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1114.592139] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9a27177-af8b-481c-a1a1-8551a3454a87 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.605810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d3ea2f-29e9-4629-ba15-f0bb701793f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.641918] env[65788]: DEBUG nova.compute.manager [req-8b972383-b0bf-40c0-b44f-508cb70e8bc5 req-2a011184-7c2c-461a-89b7-bba36772f7d5 service nova] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Detach interface failed, port_id=01a9f4ea-b2f3-4855-a441-4a5a0972a960, reason: Instance 068bb374-08a4-4f36-a9fc-bb4082f9eceb could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1114.872411] env[65788]: DEBUG oslo_vmware.api [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663124, 'name': ReconfigVM_Task, 'duration_secs': 0.225577} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.872707] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910384', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'name': 'volume-676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ec7f7a46-8dd2-4b75-866f-20e73907f1cd', 'attached_at': '', 'detached_at': '', 'volume_id': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc', 'serial': '676fc2ff-ea9f-4bd0-bd6a-86d399263acc'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1114.995825] env[65788]: DEBUG nova.scheduler.client.report [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1115.020133] env[65788]: INFO nova.compute.manager [-] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Took 1.26 seconds to deallocate network for instance. [ 1115.096345] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "6012182d-8b4e-4379-949f-d3a26211699b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.096585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "6012182d-8b4e-4379-949f-d3a26211699b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.423322] env[65788]: DEBUG nova.objects.instance [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'flavor' on Instance uuid ec7f7a46-8dd2-4b75-866f-20e73907f1cd {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.501523] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.501679] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1115.528034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.528335] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.528555] env[65788]: DEBUG nova.objects.instance [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'resources' on Instance uuid 068bb374-08a4-4f36-a9fc-bb4082f9eceb {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.599628] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1116.006928] env[65788]: DEBUG nova.compute.utils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1116.008319] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1116.008511] env[65788]: DEBUG nova.network.neutron [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1116.008831] env[65788]: WARNING neutronclient.v2_0.client [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.009170] env[65788]: WARNING neutronclient.v2_0.client [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1116.009766] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1116.010178] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1116.066459] env[65788]: DEBUG nova.policy [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1116.125194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1116.260273] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6e3fec-227f-4ec3-a67b-0509d191ce2f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.269790] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a064a03b-ec83-4019-8aa0-f03ea8e3e7bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.302633] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0ad77e-b274-452e-b688-e7dea107643e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.312473] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a42c2fd-8683-4f51-a5a4-176fc9f17e90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.328418] env[65788]: DEBUG nova.compute.provider_tree [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.408892] env[65788]: DEBUG nova.network.neutron [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Successfully created port: c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1116.432910] env[65788]: DEBUG oslo_concurrency.lockutils [None req-84903bea-19b7-42a4-8895-67d3b88f050f tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.262s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1116.519338] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1116.833621] env[65788]: DEBUG nova.scheduler.client.report [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1117.338558] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.341189] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.216s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.342897] env[65788]: INFO nova.compute.claims [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1117.364907] env[65788]: INFO nova.scheduler.client.report [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocations for instance 068bb374-08a4-4f36-a9fc-bb4082f9eceb [ 1117.486873] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.487331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.487331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.487509] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.487681] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.489935] env[65788]: INFO nova.compute.manager [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Terminating instance [ 1117.529833] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1117.555968] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1117.556247] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1117.556404] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1117.556592] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1117.556735] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1117.556880] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1117.557103] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1117.557269] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1117.557437] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1117.557614] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1117.557787] env[65788]: DEBUG nova.virt.hardware [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1117.558785] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd4d2c8-c266-4b27-ad90-f7fa25cc4aad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.567319] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a893fc66-de52-4c06-bfbd-c47c4349e0c2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.852728] env[65788]: DEBUG nova.compute.manager [req-ddf72635-e5ca-41e0-a13d-24c1742aeb0a req-a475e6ff-aee5-4f58-9185-6a84cb61f689 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Received event network-vif-plugged-c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1117.852954] env[65788]: DEBUG oslo_concurrency.lockutils [req-ddf72635-e5ca-41e0-a13d-24c1742aeb0a req-a475e6ff-aee5-4f58-9185-6a84cb61f689 service nova] Acquiring lock "27ed2344-c42b-46bc-b51a-20821f67cbf0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.853172] env[65788]: DEBUG oslo_concurrency.lockutils [req-ddf72635-e5ca-41e0-a13d-24c1742aeb0a req-a475e6ff-aee5-4f58-9185-6a84cb61f689 service nova] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.853334] env[65788]: DEBUG oslo_concurrency.lockutils [req-ddf72635-e5ca-41e0-a13d-24c1742aeb0a req-a475e6ff-aee5-4f58-9185-6a84cb61f689 service nova] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.853494] env[65788]: DEBUG nova.compute.manager [req-ddf72635-e5ca-41e0-a13d-24c1742aeb0a req-a475e6ff-aee5-4f58-9185-6a84cb61f689 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] No waiting events found dispatching network-vif-plugged-c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1117.853650] env[65788]: WARNING nova.compute.manager [req-ddf72635-e5ca-41e0-a13d-24c1742aeb0a req-a475e6ff-aee5-4f58-9185-6a84cb61f689 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Received unexpected event network-vif-plugged-c2856b60-c6f3-40b5-a301-a764e0b447eb for instance with vm_state building and task_state spawning. [ 1117.872868] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3061856c-fb49-4a51-918a-41a289c35208 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "068bb374-08a4-4f36-a9fc-bb4082f9eceb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.243s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.938395] env[65788]: DEBUG nova.network.neutron [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Successfully updated port: c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1117.994049] env[65788]: DEBUG nova.compute.manager [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1117.994399] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1117.995514] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a75d2c7-ca6a-4220-850d-33a8a3bfd4e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.004466] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1118.005268] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75be5e42-1ea5-4a43-bc39-8d5df4dbaf98 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.012067] env[65788]: DEBUG oslo_vmware.api [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1118.012067] env[65788]: value = "task-4663125" [ 1118.012067] env[65788]: _type = "Task" [ 1118.012067] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.021585] env[65788]: DEBUG oslo_vmware.api [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663125, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.442092] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-27ed2344-c42b-46bc-b51a-20821f67cbf0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.442092] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-27ed2344-c42b-46bc-b51a-20821f67cbf0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1118.442092] env[65788]: DEBUG nova.network.neutron [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1118.524871] env[65788]: DEBUG oslo_vmware.api [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663125, 'name': PowerOffVM_Task, 'duration_secs': 0.240357} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.526245] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1118.526441] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1118.527243] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4d03e9-c102-411b-9e5a-98237f0ea21c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.530119] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ad93258-a9af-4158-b453-7308aae8798a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.536743] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4467a4-ad75-4a8e-9837-ec45bcc8fa28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.569646] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7464b463-73f7-44c2-bcf2-84396b4ebded {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.578654] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71adfcd5-3586-4b31-8bb9-e658daa2df1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.593818] env[65788]: DEBUG nova.compute.provider_tree [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.617260] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1118.617524] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1118.617698] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleting the datastore file [datastore1] ec7f7a46-8dd2-4b75-866f-20e73907f1cd {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1118.617985] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f8ca1f7-023d-4134-ae07-4dfdb610d781 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.626459] env[65788]: DEBUG oslo_vmware.api [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1118.626459] env[65788]: value = "task-4663127" [ 1118.626459] env[65788]: _type = "Task" [ 1118.626459] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.636768] env[65788]: DEBUG oslo_vmware.api [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.725576] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.725813] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.944594] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1118.945020] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1118.983120] env[65788]: DEBUG nova.network.neutron [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1119.004493] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.004916] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.067871] env[65788]: WARNING neutronclient.v2_0.client [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.068578] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.068918] env[65788]: WARNING openstack [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.096877] env[65788]: DEBUG nova.scheduler.client.report [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1119.137615] env[65788]: DEBUG oslo_vmware.api [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663127, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156033} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.137912] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1119.138077] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1119.138278] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1119.138485] env[65788]: INFO nova.compute.manager [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1119.138761] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1119.138954] env[65788]: DEBUG nova.compute.manager [-] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1119.139057] env[65788]: DEBUG nova.network.neutron [-] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1119.139631] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.140209] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1119.140488] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1119.168285] env[65788]: DEBUG nova.network.neutron [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Updating instance_info_cache with network_info: [{"id": "c2856b60-c6f3-40b5-a301-a764e0b447eb", "address": "fa:16:3e:55:f8:63", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2856b60-c6", "ovs_interfaceid": "c2856b60-c6f3-40b5-a301-a764e0b447eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1119.199306] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1119.229311] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1119.602474] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1119.603030] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1119.625771] env[65788]: DEBUG nova.compute.manager [req-ab0b6630-d83a-49f2-9607-7fb178b8fb6c req-7e367f65-dae8-4f2a-ac2a-0febf615c165 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Received event network-vif-deleted-b62b03bf-7b53-4c01-880d-43cd9bf586f6 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1119.626034] env[65788]: INFO nova.compute.manager [req-ab0b6630-d83a-49f2-9607-7fb178b8fb6c req-7e367f65-dae8-4f2a-ac2a-0febf615c165 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Neutron deleted interface b62b03bf-7b53-4c01-880d-43cd9bf586f6; detaching it from the instance and deleting it from the info cache [ 1119.626156] env[65788]: DEBUG nova.network.neutron [req-ab0b6630-d83a-49f2-9607-7fb178b8fb6c req-7e367f65-dae8-4f2a-ac2a-0febf615c165 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1119.674870] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-27ed2344-c42b-46bc-b51a-20821f67cbf0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1119.675315] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Instance network_info: |[{"id": "c2856b60-c6f3-40b5-a301-a764e0b447eb", "address": "fa:16:3e:55:f8:63", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2856b60-c6", "ovs_interfaceid": "c2856b60-c6f3-40b5-a301-a764e0b447eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1119.676155] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:f8:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c2856b60-c6f3-40b5-a301-a764e0b447eb', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1119.683712] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating folder: Project (009f31c7d5bc4d369a8b96e2aa01117a). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1119.684123] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03deec0d-cfb6-4ad7-9793-31d87e08a3d2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.699121] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created folder: Project (009f31c7d5bc4d369a8b96e2aa01117a) in parent group-v910111. [ 1119.699398] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating folder: Instances. Parent ref: group-v910414. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1119.699609] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd8f2197-1d29-4a45-9d97-e89b2a6a922f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.711240] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created folder: Instances in parent group-v910414. [ 1119.711527] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1119.711743] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1119.711983] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36ae015e-5718-4ed8-98cd-2b648c69de23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.733764] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1119.733764] env[65788]: value = "task-4663130" [ 1119.733764] env[65788]: _type = "Task" [ 1119.733764] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.746027] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663130, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.760850] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.761894] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1119.763672] env[65788]: INFO nova.compute.claims [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.901088] env[65788]: DEBUG nova.compute.manager [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Received event network-changed-c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1119.901255] env[65788]: DEBUG nova.compute.manager [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Refreshing instance network info cache due to event network-changed-c2856b60-c6f3-40b5-a301-a764e0b447eb. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1119.902034] env[65788]: DEBUG oslo_concurrency.lockutils [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Acquiring lock "refresh_cache-27ed2344-c42b-46bc-b51a-20821f67cbf0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.902034] env[65788]: DEBUG oslo_concurrency.lockutils [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Acquired lock "refresh_cache-27ed2344-c42b-46bc-b51a-20821f67cbf0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1119.902214] env[65788]: DEBUG nova.network.neutron [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Refreshing network info cache for port c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1120.110174] env[65788]: DEBUG nova.compute.utils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1120.110734] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1120.111020] env[65788]: DEBUG nova.network.neutron [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1120.111914] env[65788]: WARNING neutronclient.v2_0.client [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1120.111914] env[65788]: WARNING neutronclient.v2_0.client [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1120.112535] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.113074] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.120481] env[65788]: DEBUG nova.network.neutron [-] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1120.128878] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1ca880f-fba2-4543-9486-21dccc5fd194 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.141818] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c583d88-4da6-4332-a367-91945feab1a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.177279] env[65788]: DEBUG nova.compute.manager [req-ab0b6630-d83a-49f2-9607-7fb178b8fb6c req-7e367f65-dae8-4f2a-ac2a-0febf615c165 service nova] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Detach interface failed, port_id=b62b03bf-7b53-4c01-880d-43cd9bf586f6, reason: Instance ec7f7a46-8dd2-4b75-866f-20e73907f1cd could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1120.180940] env[65788]: DEBUG nova.policy [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b220b7a1fb504d3c9b2cfd5a4b940dac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e452e2f394f549c793ecfb4f2d5cfa57', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1120.245955] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663130, 'name': CreateVM_Task, 'duration_secs': 0.332755} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.246171] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1120.246557] env[65788]: WARNING neutronclient.v2_0.client [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1120.247858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.247858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.247858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1120.247858] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67b57724-0aab-452a-852e-29cf4e1727df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.254904] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1120.254904] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a2ab0-543d-8030-e3aa-82a8d60a7c07" [ 1120.254904] env[65788]: _type = "Task" [ 1120.254904] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.272472] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a2ab0-543d-8030-e3aa-82a8d60a7c07, 'name': SearchDatastore_Task, 'duration_secs': 0.011753} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.273143] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1120.273485] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1120.274169] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.274277] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.274422] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1120.274890] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a428ff39-1ff2-41ff-9f6e-89d8a567f9c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.285757] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1120.285986] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1120.286807] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ec48f30-0f6e-4970-8cc7-5e7300eac476 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.295673] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1120.295673] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520fb1fd-e4c1-a2b1-56f6-924d8e57b85d" [ 1120.295673] env[65788]: _type = "Task" [ 1120.295673] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.306038] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520fb1fd-e4c1-a2b1-56f6-924d8e57b85d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.405033] env[65788]: WARNING neutronclient.v2_0.client [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1120.405842] env[65788]: WARNING openstack [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.406422] env[65788]: WARNING openstack [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.467543] env[65788]: DEBUG oslo_concurrency.lockutils [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.467811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.487694] env[65788]: DEBUG nova.network.neutron [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Successfully created port: dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1120.508069] env[65788]: WARNING openstack [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1120.508565] env[65788]: WARNING openstack [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1120.543640] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.544038] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.544385] env[65788]: DEBUG nova.objects.instance [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'flavor' on Instance uuid 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.623481] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1120.626410] env[65788]: INFO nova.compute.manager [-] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Took 1.49 seconds to deallocate network for instance. [ 1120.810020] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520fb1fd-e4c1-a2b1-56f6-924d8e57b85d, 'name': SearchDatastore_Task, 'duration_secs': 0.011745} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.810357] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-962c4391-bba7-4708-9aa7-d329e672fd74 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.817727] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1120.817727] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526b2f2f-eb6b-14fd-9fc4-c7e3c4f38c57" [ 1120.817727] env[65788]: _type = "Task" [ 1120.817727] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.826207] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526b2f2f-eb6b-14fd-9fc4-c7e3c4f38c57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.970527] env[65788]: INFO nova.compute.manager [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Detaching volume 7fc1ca8c-4392-42f3-9430-fec6de904bb6 [ 1120.974065] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f13063e-fe32-4cc8-a4e6-056669e11c9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.978989] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.979264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.979451] env[65788]: DEBUG nova.compute.manager [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1120.980799] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a90083a-57fa-4324-bfe4-3f3da338ef2e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.986372] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e84624-519e-46a0-b91a-5cc46a95b333 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.995328] env[65788]: DEBUG nova.compute.manager [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1120.995882] env[65788]: DEBUG nova.objects.instance [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'flavor' on Instance uuid 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.027558] env[65788]: INFO nova.virt.block_device [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Attempting to driver detach volume 7fc1ca8c-4392-42f3-9430-fec6de904bb6 from mountpoint /dev/sdb [ 1121.027780] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1121.027965] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910389', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'name': 'volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '4dcf8335-701d-4386-9de4-f14f5d375d1f', 'attached_at': '', 'detached_at': '', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'serial': '7fc1ca8c-4392-42f3-9430-fec6de904bb6'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1121.028778] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376bc506-d06c-42f4-9fff-d509e1589079 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.032277] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20c1d9b-1197-4764-aa11-85922c982cb4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.058624] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c30462-46d4-4faa-a790-11039a0b0ed5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.063769] env[65788]: WARNING neutronclient.v2_0.client [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.064430] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.064780] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.073878] env[65788]: WARNING neutronclient.v2_0.client [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.074519] env[65788]: WARNING openstack [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.074956] env[65788]: WARNING openstack [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.082801] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146709a5-6d73-4dad-a925-69dce30b4c64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.098085] env[65788]: DEBUG nova.compute.provider_tree [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.102360] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f5be6e-0096-476a-9dd3-25890a5e1318 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.134748] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2986e2d7-966b-4085-9a96-b2ae6fe8232c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.142593] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.158891] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The volume has not been displaced from its original location: [datastore2] volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6/volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1121.165110] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfiguring VM instance instance-0000004b to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1121.165486] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7854414b-3245-4f05-b6b5-a558ace9f8f0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.187335] env[65788]: DEBUG oslo_vmware.api [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1121.187335] env[65788]: value = "task-4663131" [ 1121.187335] env[65788]: _type = "Task" [ 1121.187335] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.196865] env[65788]: DEBUG oslo_vmware.api [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663131, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.236542] env[65788]: DEBUG nova.objects.instance [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'pci_requests' on Instance uuid 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.275095] env[65788]: DEBUG nova.network.neutron [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Updated VIF entry in instance network info cache for port c2856b60-c6f3-40b5-a301-a764e0b447eb. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1121.275522] env[65788]: DEBUG nova.network.neutron [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Updating instance_info_cache with network_info: [{"id": "c2856b60-c6f3-40b5-a301-a764e0b447eb", "address": "fa:16:3e:55:f8:63", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc2856b60-c6", "ovs_interfaceid": "c2856b60-c6f3-40b5-a301-a764e0b447eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1121.330648] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526b2f2f-eb6b-14fd-9fc4-c7e3c4f38c57, 'name': SearchDatastore_Task, 'duration_secs': 0.010501} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.331019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1121.331294] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 27ed2344-c42b-46bc-b51a-20821f67cbf0/27ed2344-c42b-46bc-b51a-20821f67cbf0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1121.331618] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a0f6674-c78e-42e1-bf5e-8599e112c834 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.343790] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1121.343790] env[65788]: value = "task-4663132" [ 1121.343790] env[65788]: _type = "Task" [ 1121.343790] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.352920] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663132, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.603416] env[65788]: DEBUG nova.scheduler.client.report [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1121.645352] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1121.674105] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1121.674388] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1121.674543] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1121.674724] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1121.674880] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1121.675037] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1121.675262] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1121.675425] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1121.675588] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1121.675743] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1121.675913] env[65788]: DEBUG nova.virt.hardware [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1121.676376] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.677164] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ce4742-8080-45ca-b080-bce709aaaca5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.679940] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.680150] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.680828] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.681019] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.681173] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.681319] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.681449] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1121.681589] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.687780] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2355eae-6c1e-49cc-b4b7-bd8ee71a72b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.712059] env[65788]: DEBUG oslo_vmware.api [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663131, 'name': ReconfigVM_Task, 'duration_secs': 0.244248} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.712374] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Reconfigured VM instance instance-0000004b to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1121.717667] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ccee784-b3f8-4f12-85d4-0dc3734a3d6b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.736763] env[65788]: DEBUG oslo_vmware.api [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1121.736763] env[65788]: value = "task-4663133" [ 1121.736763] env[65788]: _type = "Task" [ 1121.736763] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.740473] env[65788]: DEBUG nova.objects.base [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Object Instance<1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7> lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1121.740691] env[65788]: DEBUG nova.network.neutron [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1121.741039] env[65788]: WARNING neutronclient.v2_0.client [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.741466] env[65788]: WARNING neutronclient.v2_0.client [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1121.742269] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1121.743056] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1121.759384] env[65788]: DEBUG oslo_vmware.api [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663133, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.778958] env[65788]: DEBUG oslo_concurrency.lockutils [req-2267fd18-5804-4121-a8bf-18690769b867 req-692a79cd-950f-4e51-8e52-44e728b6a601 service nova] Releasing lock "refresh_cache-27ed2344-c42b-46bc-b51a-20821f67cbf0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1121.796599] env[65788]: DEBUG nova.policy [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1121.854704] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663132, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477464} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.855050] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 27ed2344-c42b-46bc-b51a-20821f67cbf0/27ed2344-c42b-46bc-b51a-20821f67cbf0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1121.855287] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1121.855555] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2b096ba-d0d5-4188-b828-2f81b6c9d6c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.864025] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1121.864025] env[65788]: value = "task-4663134" [ 1121.864025] env[65788]: _type = "Task" [ 1121.864025] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.874091] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663134, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.935229] env[65788]: DEBUG nova.compute.manager [req-e64260ab-6278-4dc5-8b1a-5c21d8ecc9b4 req-5d58a5d4-c8f5-4190-84b4-b957e9a65f9c service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Received event network-vif-plugged-dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1121.935229] env[65788]: DEBUG oslo_concurrency.lockutils [req-e64260ab-6278-4dc5-8b1a-5c21d8ecc9b4 req-5d58a5d4-c8f5-4190-84b4-b957e9a65f9c service nova] Acquiring lock "6012182d-8b4e-4379-949f-d3a26211699b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.935229] env[65788]: DEBUG oslo_concurrency.lockutils [req-e64260ab-6278-4dc5-8b1a-5c21d8ecc9b4 req-5d58a5d4-c8f5-4190-84b4-b957e9a65f9c service nova] Lock "6012182d-8b4e-4379-949f-d3a26211699b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.935356] env[65788]: DEBUG oslo_concurrency.lockutils [req-e64260ab-6278-4dc5-8b1a-5c21d8ecc9b4 req-5d58a5d4-c8f5-4190-84b4-b957e9a65f9c service nova] Lock "6012182d-8b4e-4379-949f-d3a26211699b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.935501] env[65788]: DEBUG nova.compute.manager [req-e64260ab-6278-4dc5-8b1a-5c21d8ecc9b4 req-5d58a5d4-c8f5-4190-84b4-b957e9a65f9c service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] No waiting events found dispatching network-vif-plugged-dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1121.935663] env[65788]: WARNING nova.compute.manager [req-e64260ab-6278-4dc5-8b1a-5c21d8ecc9b4 req-5d58a5d4-c8f5-4190-84b4-b957e9a65f9c service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Received unexpected event network-vif-plugged-dfa97bd4-874e-4c84-a6d1-d7072b745f83 for instance with vm_state building and task_state spawning. [ 1122.004038] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1122.005254] env[65788]: DEBUG nova.network.neutron [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Successfully updated port: dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1122.007365] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24aadc1d-c742-47c8-bc11-18c1f90759e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.014796] env[65788]: DEBUG oslo_vmware.api [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1122.014796] env[65788]: value = "task-4663135" [ 1122.014796] env[65788]: _type = "Task" [ 1122.014796] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.024618] env[65788]: DEBUG oslo_vmware.api [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.108857] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.109777] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1122.114377] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.971s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.114816] env[65788]: DEBUG nova.objects.instance [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'resources' on Instance uuid ec7f7a46-8dd2-4b75-866f-20e73907f1cd {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.137991] env[65788]: DEBUG nova.network.neutron [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Successfully created port: a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1122.185362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.250080] env[65788]: DEBUG oslo_vmware.api [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663133, 'name': ReconfigVM_Task, 'duration_secs': 0.169757} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.250284] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910389', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'name': 'volume-7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '4dcf8335-701d-4386-9de4-f14f5d375d1f', 'attached_at': '', 'detached_at': '', 'volume_id': '7fc1ca8c-4392-42f3-9430-fec6de904bb6', 'serial': '7fc1ca8c-4392-42f3-9430-fec6de904bb6'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1122.374393] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663134, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074954} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.374680] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1122.375542] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302c9cb4-7c68-42dc-9b61-0345b8236ec9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.398337] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 27ed2344-c42b-46bc-b51a-20821f67cbf0/27ed2344-c42b-46bc-b51a-20821f67cbf0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1122.398683] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a21dbeef-a686-42f8-bed8-ef5dd63cb8ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.419119] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1122.419119] env[65788]: value = "task-4663136" [ 1122.419119] env[65788]: _type = "Task" [ 1122.419119] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.427553] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663136, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.510431] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.510431] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1122.510608] env[65788]: DEBUG nova.network.neutron [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1122.528301] env[65788]: DEBUG oslo_vmware.api [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663135, 'name': PowerOffVM_Task, 'duration_secs': 0.210725} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.528626] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1122.528829] env[65788]: DEBUG nova.compute.manager [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1122.529720] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9433ce31-e07c-43e7-946f-421e736b065b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.618491] env[65788]: DEBUG nova.compute.utils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1122.623373] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1122.623589] env[65788]: DEBUG nova.network.neutron [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1122.623945] env[65788]: WARNING neutronclient.v2_0.client [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.624297] env[65788]: WARNING neutronclient.v2_0.client [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1122.624895] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1122.625269] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1122.672635] env[65788]: DEBUG nova.policy [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32b4adcc8fa42fba0dbffd4582fcf69', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e36859c1eb994614b2a77400f811cf50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1122.800783] env[65788]: DEBUG nova.objects.instance [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'flavor' on Instance uuid 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.817309] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd4acfd-8808-4af2-b229-bffd69d9fb92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.826703] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af45232-95d9-43cf-a7b3-6da4b39d16a5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.861443] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b3d2e6-20dd-479d-96ca-85dc61e52f0f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.871328] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e89c44b-180b-4c23-ac95-aa5f377ccc27 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.886201] env[65788]: DEBUG nova.compute.provider_tree [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1122.931189] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663136, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.977829] env[65788]: DEBUG nova.network.neutron [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Successfully created port: ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1123.013975] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.014469] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.045721] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5ffa4452-4cc9-4e43-9714-093dee5babb1 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.066s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.124722] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1123.393152] env[65788]: DEBUG nova.scheduler.client.report [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1123.431271] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663136, 'name': ReconfigVM_Task, 'duration_secs': 0.58148} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.431610] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 27ed2344-c42b-46bc-b51a-20821f67cbf0/27ed2344-c42b-46bc-b51a-20821f67cbf0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.432366] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c05a954-a6b1-4e18-a060-e15ce3f31fa2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.440596] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1123.440596] env[65788]: value = "task-4663137" [ 1123.440596] env[65788]: _type = "Task" [ 1123.440596] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.449904] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663137, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.703403] env[65788]: DEBUG nova.network.neutron [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Successfully updated port: a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1123.741426] env[65788]: DEBUG nova.network.neutron [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1123.811082] env[65788]: DEBUG oslo_concurrency.lockutils [None req-822a6774-54cf-427a-8662-89844384d86f tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.343s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.842523] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.843104] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.898592] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.784s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.904371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.718s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1123.904371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1123.904371] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1123.905256] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711945a0-1977-4236-990f-150173897a9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.921767] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044396a2-4e59-42b5-8c34-28558615816b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.930578] env[65788]: INFO nova.scheduler.client.report [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted allocations for instance ec7f7a46-8dd2-4b75-866f-20e73907f1cd [ 1123.943072] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85894690-b791-44ac-a942-e02b4adc414e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.958542] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75a566dd-eb3d-411b-a71b-f92167703e50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.961881] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663137, 'name': Rename_Task, 'duration_secs': 0.173783} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.963216] env[65788]: WARNING neutronclient.v2_0.client [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1123.963920] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1123.964313] env[65788]: WARNING openstack [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1123.972673] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1123.974060] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57abe59c-3e29-43e6-998f-ed682a5f61ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.003035] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179653MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1124.003194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.003476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.007641] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1124.007641] env[65788]: value = "task-4663138" [ 1124.007641] env[65788]: _type = "Task" [ 1124.007641] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.017320] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663138, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.067755] env[65788]: DEBUG nova.network.neutron [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Updating instance_info_cache with network_info: [{"id": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "address": "fa:16:3e:88:6a:99", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa97bd4-87", "ovs_interfaceid": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.137656] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1124.171170] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1124.171465] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1124.171636] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1124.171869] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1124.172060] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1124.172236] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1124.172554] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.172863] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1124.173412] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1124.173506] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1124.173753] env[65788]: DEBUG nova.virt.hardware [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1124.175131] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce78071-248b-4eff-b5c4-0da881ddb154 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.185845] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f18806b-f5f7-4d9b-af47-1f7a0cebd570 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.206312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.206516] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.206713] env[65788]: DEBUG nova.network.neutron [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1124.208812] env[65788]: DEBUG nova.compute.manager [req-ef3e94ca-fa4e-41ec-9422-dff5e3076cb8 req-17222273-c3ad-4c54-b2ef-fcf137d4228b service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-vif-plugged-a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1124.209892] env[65788]: DEBUG oslo_concurrency.lockutils [req-ef3e94ca-fa4e-41ec-9422-dff5e3076cb8 req-17222273-c3ad-4c54-b2ef-fcf137d4228b service nova] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1124.210123] env[65788]: DEBUG oslo_concurrency.lockutils [req-ef3e94ca-fa4e-41ec-9422-dff5e3076cb8 req-17222273-c3ad-4c54-b2ef-fcf137d4228b service nova] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.210302] env[65788]: DEBUG oslo_concurrency.lockutils [req-ef3e94ca-fa4e-41ec-9422-dff5e3076cb8 req-17222273-c3ad-4c54-b2ef-fcf137d4228b service nova] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.210535] env[65788]: DEBUG nova.compute.manager [req-ef3e94ca-fa4e-41ec-9422-dff5e3076cb8 req-17222273-c3ad-4c54-b2ef-fcf137d4228b service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] No waiting events found dispatching network-vif-plugged-a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1124.210686] env[65788]: WARNING nova.compute.manager [req-ef3e94ca-fa4e-41ec-9422-dff5e3076cb8 req-17222273-c3ad-4c54-b2ef-fcf137d4228b service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received unexpected event network-vif-plugged-a99b6060-25f2-4e56-bc5b-2be77fe833dc for instance with vm_state active and task_state None. [ 1124.369377] env[65788]: DEBUG nova.compute.manager [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Received event network-changed-dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1124.369577] env[65788]: DEBUG nova.compute.manager [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Refreshing instance network info cache due to event network-changed-dfa97bd4-874e-4c84-a6d1-d7072b745f83. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1124.369759] env[65788]: DEBUG oslo_concurrency.lockutils [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Acquiring lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.449761] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a179a7b5-73c7-4226-98bc-9cc95d0309ea tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ec7f7a46-8dd2-4b75-866f-20e73907f1cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.963s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.530834] env[65788]: DEBUG oslo_vmware.api [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663138, 'name': PowerOnVM_Task, 'duration_secs': 0.478556} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.531583] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1124.531802] env[65788]: INFO nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Took 7.00 seconds to spawn the instance on the hypervisor. [ 1124.531976] env[65788]: DEBUG nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1124.533282] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6d7569-7d9a-4ec5-9260-715c987b9c76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.570920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1124.571352] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Instance network_info: |[{"id": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "address": "fa:16:3e:88:6a:99", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa97bd4-87", "ovs_interfaceid": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1124.571812] env[65788]: DEBUG oslo_concurrency.lockutils [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Acquired lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.571901] env[65788]: DEBUG nova.network.neutron [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Refreshing network info cache for port dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1124.574038] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:6a:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56398cc0-e39f-410f-8036-8c2a6870e26f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfa97bd4-874e-4c84-a6d1-d7072b745f83', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.581304] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Creating folder: Project (e452e2f394f549c793ecfb4f2d5cfa57). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1124.582725] env[65788]: WARNING neutronclient.v2_0.client [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.583365] env[65788]: WARNING openstack [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.583715] env[65788]: WARNING openstack [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.598897] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1deb97ef-262f-4cb5-ace7-2d32d6c8112f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.608661] env[65788]: DEBUG nova.network.neutron [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Successfully updated port: ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1124.621751] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Created folder: Project (e452e2f394f549c793ecfb4f2d5cfa57) in parent group-v910111. [ 1124.622013] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Creating folder: Instances. Parent ref: group-v910417. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1124.622291] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2ee7397-3fe5-404c-b462-ec41c05d2172 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.634580] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Created folder: Instances in parent group-v910417. [ 1124.636411] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1124.636411] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1124.636411] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1415658-6077-452a-8a79-be535c568d75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.656674] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.656674] env[65788]: value = "task-4663141" [ 1124.656674] env[65788]: _type = "Task" [ 1124.656674] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.664907] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663141, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.713779] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.714095] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.777292] env[65788]: WARNING openstack [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.777684] env[65788]: WARNING openstack [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.801198] env[65788]: WARNING nova.network.neutron [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1124.836647] env[65788]: WARNING neutronclient.v2_0.client [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.837499] env[65788]: WARNING openstack [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.837945] env[65788]: WARNING openstack [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.853137] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.853660] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1124.934069] env[65788]: DEBUG nova.objects.instance [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'flavor' on Instance uuid 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.971155] env[65788]: DEBUG nova.network.neutron [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Updated VIF entry in instance network info cache for port dfa97bd4-874e-4c84-a6d1-d7072b745f83. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1124.971518] env[65788]: DEBUG nova.network.neutron [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Updating instance_info_cache with network_info: [{"id": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "address": "fa:16:3e:88:6a:99", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa97bd4-87", "ovs_interfaceid": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1124.996350] env[65788]: WARNING neutronclient.v2_0.client [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1124.997108] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1124.997616] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.023815] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.023815] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.023815] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.024328] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.024328] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.028087] env[65788]: INFO nova.compute.manager [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Terminating instance [ 1125.053931] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.055890] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1a701758-a056-4948-9069-2a7168f2cc8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.056363] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9d2f0d5e-793d-4790-ad31-be3d372be4fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.056516] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4dcf8335-701d-4386-9de4-f14f5d375d1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.056633] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.056745] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.056856] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 18be7322-3359-49c9-a181-f9228d5c16d7 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.056962] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance b9d8a32e-1f0f-405c-9543-9e53046d248e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.057078] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9946cef4-b8c7-41b4-9299-eace1929bf3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.057185] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 27ed2344-c42b-46bc-b51a-20821f67cbf0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.057312] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6012182d-8b4e-4379-949f-d3a26211699b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.057428] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 5621bbe8-131f-4266-9109-a3be445310a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1125.057628] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1125.057775] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=100GB used_disk=11GB total_vcpus=48 used_vcpus=12 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '12', 'num_vm_active': '8', 'num_task_None': '9', 'num_os_type_None': '12', 'num_proj_ffbe44d63c1d432e97849f15615329e1': '3', 'io_workload': '3', 'num_proj_94977b75768444138fb914cabddc57c5': '1', 'num_proj_828bafa0635a42c7be55b1fb367e1734': '1', 'num_proj_676963c25a3741998b10790689242dbb': '1', 'num_vm_stopped': '1', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '1', 'num_proj_2b70d3c8627449eaa6372ebe3bd90233': '1', 'num_proj_021bf6dd87e340969a57e7e5b33268e7': '1', 'num_vm_building': '3', 'num_task_spawning': '3', 'num_proj_009f31c7d5bc4d369a8b96e2aa01117a': '1', 'num_proj_e452e2f394f549c793ecfb4f2d5cfa57': '1', 'num_proj_e36859c1eb994614b2a77400f811cf50': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1125.061414] env[65788]: INFO nova.compute.manager [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Took 13.89 seconds to build instance. [ 1125.114254] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.114254] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.114254] env[65788]: DEBUG nova.network.neutron [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1125.144986] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.146179] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.168172] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663141, 'name': CreateVM_Task, 'duration_secs': 0.409815} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.168526] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1125.168858] env[65788]: WARNING neutronclient.v2_0.client [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.169279] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.169450] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.169809] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1125.172801] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-146d7fb8-bde6-4839-9636-82b5959f4685 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.179996] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1125.179996] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52368d01-278d-56c1-b7a0-886b71a29eb5" [ 1125.179996] env[65788]: _type = "Task" [ 1125.179996] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.189436] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52368d01-278d-56c1-b7a0-886b71a29eb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.259344] env[65788]: WARNING neutronclient.v2_0.client [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.260478] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.260869] env[65788]: WARNING openstack [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.283101] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce44061-1e4c-42f5-8e41-c9bec4c5498b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.291968] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40731407-fc90-4bfb-b97e-a7860b15a1ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.331743] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730a7ffd-d553-4fff-bbc8-fb781ecd26cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.340904] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cddf0f41-a66f-41fd-9c03-8ceadc6abdd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.355208] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.382780] env[65788]: DEBUG nova.network.neutron [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a99b6060-25f2-4e56-bc5b-2be77fe833dc", "address": "fa:16:3e:df:81:a2", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa99b6060-25", "ovs_interfaceid": "a99b6060-25f2-4e56-bc5b-2be77fe833dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1125.444598] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.444794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.445150] env[65788]: DEBUG nova.network.neutron [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1125.445374] env[65788]: DEBUG nova.objects.instance [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'info_cache' on Instance uuid 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.474552] env[65788]: DEBUG oslo_concurrency.lockutils [req-8836337f-8f24-4800-8ab5-f969a2761838 req-2ef47561-ea0c-4958-b002-a9d2c32c491f service nova] Releasing lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.532880] env[65788]: DEBUG nova.compute.manager [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1125.533121] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.534085] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7128c1f5-eb85-40ad-ab3a-55b5cd92492f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.546456] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.546742] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-930f1d0e-77c9-429a-a192-fff8c88ba3b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.554190] env[65788]: DEBUG oslo_vmware.api [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1125.554190] env[65788]: value = "task-4663142" [ 1125.554190] env[65788]: _type = "Task" [ 1125.554190] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.563572] env[65788]: DEBUG oslo_vmware.api [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.564042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f99ee299-8189-4d7c-ac9b-35cdb6186f10 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.400s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.620664] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.621549] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.665519] env[65788]: DEBUG nova.network.neutron [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1125.686521] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.686988] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.701171] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52368d01-278d-56c1-b7a0-886b71a29eb5, 'name': SearchDatastore_Task, 'duration_secs': 0.013063} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.701415] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.701650] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.701903] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.702061] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.702265] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.703124] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff8cb699-6bf7-4ab1-90db-d55ba7acbf58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.714798] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.715214] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1125.715995] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9248119-9ca4-462d-a99e-6ea568b36189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.724176] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1125.724176] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ec455b-3d88-97f2-856e-b890cc17a100" [ 1125.724176] env[65788]: _type = "Task" [ 1125.724176] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.739086] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ec455b-3d88-97f2-856e-b890cc17a100, 'name': SearchDatastore_Task, 'duration_secs': 0.011841} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.739962] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4042fc61-e2e3-4ef3-87cb-566af6145690 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.746440] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1125.746440] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52032ac2-a09a-6d60-3298-45af0440ad5e" [ 1125.746440] env[65788]: _type = "Task" [ 1125.746440] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.756214] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52032ac2-a09a-6d60-3298-45af0440ad5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.759705] env[65788]: WARNING neutronclient.v2_0.client [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1125.760564] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1125.760694] env[65788]: WARNING openstack [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1125.847448] env[65788]: DEBUG nova.network.neutron [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [{"id": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "address": "fa:16:3e:3d:6c:91", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1870b5-8b", "ovs_interfaceid": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1125.858959] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1125.885607] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.886485] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.886673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.887563] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbbba19-aae6-45a4-890d-eb39c5ab7024 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.906807] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1125.907353] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1125.907526] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1125.907747] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1125.907883] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1125.907988] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1125.908217] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1125.908382] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1125.908550] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1125.908714] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1125.908891] env[65788]: DEBUG nova.virt.hardware [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1125.915207] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Reconfiguring VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1125.915952] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8d7304a-c4ae-4970-adef-4e8ade91333f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.933952] env[65788]: DEBUG oslo_vmware.api [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1125.933952] env[65788]: value = "task-4663143" [ 1125.933952] env[65788]: _type = "Task" [ 1125.933952] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.942547] env[65788]: DEBUG oslo_vmware.api [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663143, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.948956] env[65788]: DEBUG nova.objects.base [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Object Instance<475291c8-a16d-4980-9ef4-7ed6a4dc382e> lazy-loaded attributes: flavor,info_cache {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1126.068403] env[65788]: DEBUG oslo_vmware.api [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663142, 'name': PowerOffVM_Task, 'duration_secs': 0.214142} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.068775] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.069033] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.069382] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e5365f5-51c9-4e96-bb06-a561231702ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.140963] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.141240] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.141531] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleting the datastore file [datastore2] 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.141998] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3708d9bc-7b1e-4311-8c8c-2e51740e0ff0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.150639] env[65788]: DEBUG oslo_vmware.api [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1126.150639] env[65788]: value = "task-4663145" [ 1126.150639] env[65788]: _type = "Task" [ 1126.150639] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.160954] env[65788]: DEBUG oslo_vmware.api [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.257791] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52032ac2-a09a-6d60-3298-45af0440ad5e, 'name': SearchDatastore_Task, 'duration_secs': 0.012153} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.257791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1126.257973] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/6012182d-8b4e-4379-949f-d3a26211699b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1126.258184] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba4b81bc-a532-4d53-afb2-dd46c97f7d5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.266585] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1126.266585] env[65788]: value = "task-4663146" [ 1126.266585] env[65788]: _type = "Task" [ 1126.266585] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.278568] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663146, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.352456] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1126.352865] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Instance network_info: |[{"id": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "address": "fa:16:3e:3d:6c:91", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1870b5-8b", "ovs_interfaceid": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1126.353388] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:6c:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1d468f87-964a-4fb6-bab3-b83f6f2646b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec1870b5-8b77-4ad8-9779-fdcd725dd800', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1126.361278] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1126.361541] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1126.361793] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56f77b40-9418-4eab-a4f5-183202745589 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.376524] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1126.376805] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.373s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.383404] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1126.383404] env[65788]: value = "task-4663147" [ 1126.383404] env[65788]: _type = "Task" [ 1126.383404] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.392445] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663147, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.446408] env[65788]: DEBUG oslo_vmware.api [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663143, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.451953] env[65788]: WARNING neutronclient.v2_0.client [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.451953] env[65788]: WARNING openstack [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.452325] env[65788]: WARNING openstack [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.568357] env[65788]: DEBUG nova.compute.manager [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-changed-a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1126.568357] env[65788]: DEBUG nova.compute.manager [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Refreshing instance network info cache due to event network-changed-a99b6060-25f2-4e56-bc5b-2be77fe833dc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1126.569443] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Acquiring lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.569443] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Acquired lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.569443] env[65788]: DEBUG nova.network.neutron [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Refreshing network info cache for port a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1126.581345] env[65788]: WARNING openstack [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.581770] env[65788]: WARNING openstack [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.639038] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1126.657361] env[65788]: WARNING neutronclient.v2_0.client [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.658148] env[65788]: WARNING openstack [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.658587] env[65788]: WARNING openstack [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.678036] env[65788]: DEBUG oslo_vmware.api [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199271} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.678389] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.678568] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.678727] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1126.678904] env[65788]: INFO nova.compute.manager [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1126.679254] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1126.679589] env[65788]: DEBUG nova.compute.manager [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1126.679589] env[65788]: DEBUG nova.network.neutron [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1126.679833] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.680540] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1126.680897] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1126.748780] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.767484] env[65788]: DEBUG nova.network.neutron [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1126.782242] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663146, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.820080] env[65788]: DEBUG nova.compute.manager [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Received event network-vif-plugged-ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1126.820558] env[65788]: DEBUG oslo_concurrency.lockutils [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.820558] env[65788]: DEBUG oslo_concurrency.lockutils [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Lock "5621bbe8-131f-4266-9109-a3be445310a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.820702] env[65788]: DEBUG oslo_concurrency.lockutils [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Lock "5621bbe8-131f-4266-9109-a3be445310a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.821302] env[65788]: DEBUG nova.compute.manager [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] No waiting events found dispatching network-vif-plugged-ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1126.821302] env[65788]: WARNING nova.compute.manager [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Received unexpected event network-vif-plugged-ec1870b5-8b77-4ad8-9779-fdcd725dd800 for instance with vm_state building and task_state spawning. [ 1126.821582] env[65788]: DEBUG nova.compute.manager [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Received event network-changed-ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1126.821631] env[65788]: DEBUG nova.compute.manager [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Refreshing instance network info cache due to event network-changed-ec1870b5-8b77-4ad8-9779-fdcd725dd800. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1126.822718] env[65788]: DEBUG oslo_concurrency.lockutils [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Acquiring lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.822718] env[65788]: DEBUG oslo_concurrency.lockutils [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Acquired lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.822718] env[65788]: DEBUG nova.network.neutron [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Refreshing network info cache for port ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1126.893795] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663147, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.904147] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.904147] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.945668] env[65788]: DEBUG oslo_vmware.api [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663143, 'name': ReconfigVM_Task, 'duration_secs': 0.887527} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.946112] env[65788]: WARNING neutronclient.v2_0.client [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1126.946397] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1126.946623] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Reconfigured VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1127.072748] env[65788]: WARNING neutronclient.v2_0.client [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.073467] env[65788]: WARNING openstack [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.073823] env[65788]: WARNING openstack [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.201780] env[65788]: WARNING openstack [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.202455] env[65788]: WARNING openstack [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.275178] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.281727] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663146, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.613691} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.282146] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/6012182d-8b4e-4379-949f-d3a26211699b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1127.284699] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1127.284699] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a370ff5e-ddab-406c-8e65-c6a18fdd218c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.291340] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1127.291340] env[65788]: value = "task-4663148" [ 1127.291340] env[65788]: _type = "Task" [ 1127.291340] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.302916] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663148, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.312188] env[65788]: WARNING neutronclient.v2_0.client [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.312974] env[65788]: WARNING openstack [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.313362] env[65788]: WARNING openstack [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.326490] env[65788]: WARNING neutronclient.v2_0.client [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.328029] env[65788]: WARNING openstack [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.328503] env[65788]: WARNING openstack [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.340504] env[65788]: DEBUG nova.network.neutron [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1127.399985] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663147, 'name': CreateVM_Task, 'duration_secs': 0.548833} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.399985] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1127.400465] env[65788]: WARNING neutronclient.v2_0.client [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.400857] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.401045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.401417] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1127.401701] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6494b5b0-c348-4ce3-bf56-e4c78ffd797c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.408317] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1127.411743] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1127.411743] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f7bd7d-af73-d0bd-bea0-eb145bd8878b" [ 1127.411743] env[65788]: _type = "Task" [ 1127.411743] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.421393] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f7bd7d-af73-d0bd-bea0-eb145bd8878b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.435030] env[65788]: DEBUG nova.network.neutron [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updated VIF entry in instance network info cache for port a99b6060-25f2-4e56-bc5b-2be77fe833dc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1127.435638] env[65788]: DEBUG nova.network.neutron [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a99b6060-25f2-4e56-bc5b-2be77fe833dc", "address": "fa:16:3e:df:81:a2", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa99b6060-25", "ovs_interfaceid": "a99b6060-25f2-4e56-bc5b-2be77fe833dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1127.454514] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c557f15f-c524-4ced-b7f1-b5549ff98ad1 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.909s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1127.471019] env[65788]: WARNING openstack [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.471019] env[65788]: WARNING openstack [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.539265] env[65788]: WARNING neutronclient.v2_0.client [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1127.540127] env[65788]: WARNING openstack [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1127.540523] env[65788]: WARNING openstack [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1127.627017] env[65788]: DEBUG nova.network.neutron [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updated VIF entry in instance network info cache for port ec1870b5-8b77-4ad8-9779-fdcd725dd800. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1127.629122] env[65788]: DEBUG nova.network.neutron [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [{"id": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "address": "fa:16:3e:3d:6c:91", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1870b5-8b", "ovs_interfaceid": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1127.799331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "84a4e3bf-76cf-4699-8996-230c227cc1f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.799653] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.807476] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663148, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124946} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.807819] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.808620] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e39e8e-9d15-4fe5-a53f-ecad6cd5aec8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.832817] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/6012182d-8b4e-4379-949f-d3a26211699b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.833508] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a736c8f1-3cb8-4f7b-be21-370b786d0a29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.848584] env[65788]: INFO nova.compute.manager [-] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Took 1.17 seconds to deallocate network for instance. [ 1127.856701] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1127.856701] env[65788]: value = "task-4663149" [ 1127.856701] env[65788]: _type = "Task" [ 1127.856701] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.868457] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663149, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.926212] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f7bd7d-af73-d0bd-bea0-eb145bd8878b, 'name': SearchDatastore_Task, 'duration_secs': 0.021476} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.926534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.926778] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1127.927027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.927593] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1127.927593] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1127.927760] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51f1bcbd-d845-4bcb-9107-bc005172bb73 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.933709] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1127.934294] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1127.935614] env[65788]: INFO nova.compute.claims [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1127.939137] env[65788]: DEBUG oslo_concurrency.lockutils [req-bc9e36a4-98c0-4173-8b29-c4f7dc03815a req-a2c13d65-4c83-4c29-9a5f-aac6576bc431 service nova] Releasing lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.941490] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1127.941490] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1127.942466] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc6c4744-bb0c-4875-ad20-c91c592a273c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.948779] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1127.948779] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc2a63-9a3c-6c2a-c364-965907b49371" [ 1127.948779] env[65788]: _type = "Task" [ 1127.948779] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.959854] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc2a63-9a3c-6c2a-c364-965907b49371, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.130908] env[65788]: DEBUG oslo_concurrency.lockutils [req-90f42bb6-fa28-4279-b2e8-2175dddcf9ff req-b1427448-6330-477b-84c2-26a0484aac20 service nova] Releasing lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1128.280697] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1128.281184] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dbc5cb3-1f54-41d6-864b-e1b312d6d8ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.289450] env[65788]: DEBUG oslo_vmware.api [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1128.289450] env[65788]: value = "task-4663150" [ 1128.289450] env[65788]: _type = "Task" [ 1128.289450] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.299897] env[65788]: DEBUG oslo_vmware.api [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663150, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.302617] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1128.355960] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.369191] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663149, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.462030] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc2a63-9a3c-6c2a-c364-965907b49371, 'name': SearchDatastore_Task, 'duration_secs': 0.016514} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.463540] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4db438af-5c7a-49a0-8457-9e33d31737ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.471435] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1128.471435] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52117733-92d6-ebf7-195e-038ceb1f3e74" [ 1128.471435] env[65788]: _type = "Task" [ 1128.471435] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.484089] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52117733-92d6-ebf7-195e-038ceb1f3e74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.604498] env[65788]: DEBUG nova.compute.manager [req-9285aaf6-3658-4dda-86c0-dd42b4804d8e req-59dd74a5-8156-45e1-ae3f-43d3a84003a6 service nova] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Received event network-vif-deleted-eba2c4b7-13de-4fe3-95fb-86d9494cd97a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1128.801430] env[65788]: DEBUG oslo_vmware.api [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663150, 'name': PowerOnVM_Task, 'duration_secs': 0.432153} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.801430] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.801920] env[65788]: DEBUG nova.compute.manager [None req-7cdfb52b-af1a-4201-a8f0-f776ba3762ae tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1128.802552] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f84f0c-8c1c-4695-a5fd-10ef3bec6f68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.830130] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.868950] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663149, 'name': ReconfigVM_Task, 'duration_secs': 0.547617} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.869260] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/6012182d-8b4e-4379-949f-d3a26211699b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.869906] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3d49eca-d85b-451a-aafb-9d81e80d8ff1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.877401] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1128.877401] env[65788]: value = "task-4663151" [ 1128.877401] env[65788]: _type = "Task" [ 1128.877401] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.886228] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663151, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.982797] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52117733-92d6-ebf7-195e-038ceb1f3e74, 'name': SearchDatastore_Task, 'duration_secs': 0.013229} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.985349] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1128.985620] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0/5621bbe8-131f-4266-9109-a3be445310a0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1128.986105] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e39c0157-4af1-4d80-830f-b5597a73aa17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.995216] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1128.995216] env[65788]: value = "task-4663152" [ 1128.995216] env[65788]: _type = "Task" [ 1128.995216] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.006878] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.015748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-a99b6060-25f2-4e56-bc5b-2be77fe833dc" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.016049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-a99b6060-25f2-4e56-bc5b-2be77fe833dc" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.140773] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d4589d-fe4c-47f2-b5bb-0d97f2622be2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.144319] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.144645] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.150149] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751690b6-29d4-424c-bba3-8a50cb783e5b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.183531] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3de7af-8bcd-4a5a-9ff8-0611376a4ee3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.192040] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0d9944-d24b-46b8-b8f1-83f7d5a0f665 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.207040] env[65788]: DEBUG nova.compute.provider_tree [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1129.388911] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663151, 'name': Rename_Task, 'duration_secs': 0.186917} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.389286] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1129.389607] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bfb3550b-105e-4b81-b374-89b0b6e4742b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.398122] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1129.398122] env[65788]: value = "task-4663153" [ 1129.398122] env[65788]: _type = "Task" [ 1129.398122] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.408738] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663153, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.511210] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663152, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.519199] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.519436] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.520462] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3ddd55-0c80-42fc-91a1-9a4e65722b9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.545705] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29df293-2d02-4390-a3e2-7bdcb36c523e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.570862] env[65788]: WARNING neutronclient.v2_0.client [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1129.577168] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Reconfiguring VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1129.577668] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7aebf891-926d-4650-b7fb-dd23e5e1db64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.599976] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1129.599976] env[65788]: value = "task-4663154" [ 1129.599976] env[65788]: _type = "Task" [ 1129.599976] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.610298] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.733040] env[65788]: ERROR nova.scheduler.client.report [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [req-415ca8f9-4694-4e69-9779-81ae7b2bb642] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-415ca8f9-4694-4e69-9779-81ae7b2bb642"}]} [ 1129.753932] env[65788]: DEBUG nova.scheduler.client.report [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1129.771414] env[65788]: DEBUG nova.scheduler.client.report [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1129.771669] env[65788]: DEBUG nova.compute.provider_tree [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1129.786384] env[65788]: DEBUG nova.scheduler.client.report [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1129.810117] env[65788]: DEBUG nova.scheduler.client.report [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1129.910621] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663153, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.003985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346e476f-2b35-4d4c-965a-58ee2a9a075f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.014839] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066320fc-29ad-4520-a381-dec42c4bad55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.018008] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.928278} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.018283] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0/5621bbe8-131f-4266-9109-a3be445310a0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1130.018499] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1130.019064] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fca5bfb0-0219-4d70-b302-717cccb5d5d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.050933] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325ce56b-4486-4627-9820-0f092a808cb1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.056870] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1130.056870] env[65788]: value = "task-4663155" [ 1130.056870] env[65788]: _type = "Task" [ 1130.056870] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.064090] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e41681-d1ad-49be-80cd-dc3b989951fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.074368] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663155, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.083557] env[65788]: DEBUG nova.compute.provider_tree [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1130.111887] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.408979] env[65788]: DEBUG oslo_vmware.api [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663153, 'name': PowerOnVM_Task, 'duration_secs': 0.782185} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.409328] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1130.409589] env[65788]: INFO nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Took 8.76 seconds to spawn the instance on the hypervisor. [ 1130.409809] env[65788]: DEBUG nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1130.410625] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cf353c-b80b-47df-90a3-0ae539edf4d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.569328] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663155, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092413} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.569472] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1130.570329] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcedf6d-cb6b-4973-a3f0-8822930da011 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.595182] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0/5621bbe8-131f-4266-9109-a3be445310a0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1130.597149] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae67fedd-c262-4dbe-be49-3129bea41aae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.622645] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.624413] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1130.624413] env[65788]: value = "task-4663156" [ 1130.624413] env[65788]: _type = "Task" [ 1130.624413] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.633299] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663156, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.636718] env[65788]: DEBUG nova.scheduler.client.report [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1130.637040] env[65788]: DEBUG nova.compute.provider_tree [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 149 to 150 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1130.637288] env[65788]: DEBUG nova.compute.provider_tree [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1130.640990] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.641212] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.927577] env[65788]: INFO nova.compute.manager [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Took 14.82 seconds to build instance. [ 1131.123552] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.134812] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663156, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.143817] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.210s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.144416] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1131.147537] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.793s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.147765] env[65788]: DEBUG nova.objects.instance [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'resources' on Instance uuid 4dcf8335-701d-4386-9de4-f14f5d375d1f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.429253] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6a5e0813-b4af-4203-90d9-520bf0bf86d5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "6012182d-8b4e-4379-949f-d3a26211699b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.332s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.622986] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.634832] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663156, 'name': ReconfigVM_Task, 'duration_secs': 0.728933} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.635170] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0/5621bbe8-131f-4266-9109-a3be445310a0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.635809] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-414a682a-92f3-460e-adc3-033eee49b8d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.637625] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.643576] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1131.643576] env[65788]: value = "task-4663157" [ 1131.643576] env[65788]: _type = "Task" [ 1131.643576] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.650401] env[65788]: DEBUG nova.compute.utils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1131.655361] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1131.655568] env[65788]: DEBUG nova.network.neutron [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1131.655891] env[65788]: WARNING neutronclient.v2_0.client [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.656203] env[65788]: WARNING neutronclient.v2_0.client [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1131.656787] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1131.657151] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1131.665894] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663157, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.728053] env[65788]: DEBUG nova.policy [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca13693271604f5790bdc8e19aa208a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7aa1dec53aa4436daa0f3223b92f9b79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1131.840882] env[65788]: INFO nova.compute.manager [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Rescuing [ 1131.841174] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.841333] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.841497] env[65788]: DEBUG nova.network.neutron [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1131.864169] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3ad2fe-3595-43f6-b843-170f828d648e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.873524] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470b7752-cec5-420e-b00e-01d3cab4524a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.907515] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5964af3e-e39b-4a2c-8258-0bd49330321e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.916044] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8de0067-cefa-4846-8c52-1634baedc664 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.930435] env[65788]: DEBUG nova.compute.provider_tree [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.037857] env[65788]: DEBUG nova.network.neutron [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Successfully created port: 731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1132.125642] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.154851] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663157, 'name': Rename_Task, 'duration_secs': 0.245407} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.155189] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1132.155772] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b924857-0cef-4482-bbfa-f2709fcc669b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.162791] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1132.162791] env[65788]: value = "task-4663158" [ 1132.162791] env[65788]: _type = "Task" [ 1132.162791] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.167016] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1132.172914] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.345080] env[65788]: WARNING neutronclient.v2_0.client [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.345080] env[65788]: WARNING openstack [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.345336] env[65788]: WARNING openstack [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.434337] env[65788]: DEBUG nova.scheduler.client.report [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1132.449338] env[65788]: WARNING openstack [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.449960] env[65788]: WARNING openstack [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.546646] env[65788]: WARNING neutronclient.v2_0.client [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1132.547446] env[65788]: WARNING openstack [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1132.547804] env[65788]: WARNING openstack [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1132.629986] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.632696] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.637615] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.637890] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1132.651649] env[65788]: DEBUG nova.network.neutron [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Updating instance_info_cache with network_info: [{"id": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "address": "fa:16:3e:88:6a:99", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa97bd4-87", "ovs_interfaceid": "dfa97bd4-874e-4c84-a6d1-d7072b745f83", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1132.678874] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663158, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.941363] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1132.944049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.114s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1132.946063] env[65788]: INFO nova.compute.claims [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1132.972392] env[65788]: INFO nova.scheduler.client.report [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted allocations for instance 4dcf8335-701d-4386-9de4-f14f5d375d1f [ 1133.126025] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.155250] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "refresh_cache-6012182d-8b4e-4379-949f-d3a26211699b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.173276] env[65788]: DEBUG oslo_vmware.api [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663158, 'name': PowerOnVM_Task, 'duration_secs': 0.53395} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.173711] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1133.173991] env[65788]: INFO nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Took 9.04 seconds to spawn the instance on the hypervisor. [ 1133.174265] env[65788]: DEBUG nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1133.175127] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e84e07-3306-4c14-91b2-ff01b04015ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.180292] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1133.203534] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1133.203838] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1133.204120] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1133.204417] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1133.204640] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1133.205256] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1133.205256] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.205256] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1133.205464] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1133.205628] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1133.205815] env[65788]: DEBUG nova.virt.hardware [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1133.206976] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9510ac7-f002-4e26-b282-a02acc2db90c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.217182] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc9d178-970c-4864-962d-ffef3904d5e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.482830] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3667851f-3181-4f47-a0e0-f3629fe52b1e tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "4dcf8335-701d-4386-9de4-f14f5d375d1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.459s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.485719] env[65788]: DEBUG nova.compute.manager [req-561a8bdd-c69b-4f4b-854d-ffca0a26e731 req-a940f522-8b83-4094-a403-aa656da83b1b service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Received event network-vif-plugged-731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1133.486392] env[65788]: DEBUG oslo_concurrency.lockutils [req-561a8bdd-c69b-4f4b-854d-ffca0a26e731 req-a940f522-8b83-4094-a403-aa656da83b1b service nova] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.486792] env[65788]: DEBUG oslo_concurrency.lockutils [req-561a8bdd-c69b-4f4b-854d-ffca0a26e731 req-a940f522-8b83-4094-a403-aa656da83b1b service nova] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.486983] env[65788]: DEBUG oslo_concurrency.lockutils [req-561a8bdd-c69b-4f4b-854d-ffca0a26e731 req-a940f522-8b83-4094-a403-aa656da83b1b service nova] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.487216] env[65788]: DEBUG nova.compute.manager [req-561a8bdd-c69b-4f4b-854d-ffca0a26e731 req-a940f522-8b83-4094-a403-aa656da83b1b service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] No waiting events found dispatching network-vif-plugged-731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1133.487383] env[65788]: WARNING nova.compute.manager [req-561a8bdd-c69b-4f4b-854d-ffca0a26e731 req-a940f522-8b83-4094-a403-aa656da83b1b service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Received unexpected event network-vif-plugged-731c93dc-1d06-4943-bac0-fc800f9f449d for instance with vm_state building and task_state spawning. [ 1133.580512] env[65788]: DEBUG nova.network.neutron [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Successfully updated port: 731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1133.628142] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.700211] env[65788]: INFO nova.compute.manager [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Took 13.96 seconds to build instance. [ 1134.082619] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.082887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.083099] env[65788]: DEBUG nova.network.neutron [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1134.133063] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.156295] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb1ef22-50d3-4611-b0dc-54f357739e91 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.166396] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e4e597-5a59-4b8d-9324-6d3e8422c52b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.202807] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ffb59086-3acd-4c36-b76c-4a20063f3161 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.477s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.203688] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b73c19-c05d-4a98-b2e8-462fb90c415b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.214425] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430c423f-89d8-46f8-8eab-7f230166b6f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.230814] env[65788]: DEBUG nova.compute.provider_tree [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.587627] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.587969] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.630864] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.632054] env[65788]: DEBUG nova.network.neutron [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1134.637719] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.677197] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.677706] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.710779] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1134.712931] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29fbd634-cc63-4b7a-817b-2b18d2ef91dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.723096] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1134.723096] env[65788]: value = "task-4663159" [ 1134.723096] env[65788]: _type = "Task" [ 1134.723096] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.732356] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.734425] env[65788]: DEBUG nova.scheduler.client.report [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1134.759274] env[65788]: WARNING neutronclient.v2_0.client [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1134.759980] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1134.760409] env[65788]: WARNING openstack [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1134.893531] env[65788]: DEBUG nova.network.neutron [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updating instance_info_cache with network_info: [{"id": "731c93dc-1d06-4943-bac0-fc800f9f449d", "address": "fa:16:3e:bb:7d:4f", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap731c93dc-1d", "ovs_interfaceid": "731c93dc-1d06-4943-bac0-fc800f9f449d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1135.137475] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.140585] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.234965] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663159, 'name': PowerOffVM_Task, 'duration_secs': 0.465317} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.235318] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.236204] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4af347-2991-4e1c-8cb7-48d964d8c95b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.256634] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.257192] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1135.260763] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.120s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.260763] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.260763] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1135.261457] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f72b39-42e5-4484-a2a4-c1d4686b293e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.265502] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6649fe25-6f47-4cbd-bca8-8b3e40bb3c9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.280116] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb1e4b1-d363-434b-9f53-b6e792869d17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.298884] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b39f528f-cab2-4d7f-b234-258380a6943c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.312353] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.313452] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08f10cb-3081-4f0f-9f5a-9bc65e20dca4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.317323] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3400f47a-953f-4c16-9c28-16acd708ae7e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.352034] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179653MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1135.352258] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.352534] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.356522] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1135.356522] env[65788]: value = "task-4663160" [ 1135.356522] env[65788]: _type = "Task" [ 1135.356522] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.367074] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1135.367322] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1135.367572] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.367716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.367896] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1135.368550] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd169abc-bc10-469d-8c3d-4a3acafb88cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.380256] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1135.380444] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1135.381241] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b74c3e6-6d4e-4666-9cf2-d1325d210426 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.387359] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1135.387359] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520350ca-a4b3-24cc-f179-82c2a81c4079" [ 1135.387359] env[65788]: _type = "Task" [ 1135.387359] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.397285] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.397654] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Instance network_info: |[{"id": "731c93dc-1d06-4943-bac0-fc800f9f449d", "address": "fa:16:3e:bb:7d:4f", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap731c93dc-1d", "ovs_interfaceid": "731c93dc-1d06-4943-bac0-fc800f9f449d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1135.398483] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520350ca-a4b3-24cc-f179-82c2a81c4079, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.398856] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:7d:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '731c93dc-1d06-4943-bac0-fc800f9f449d', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1135.406631] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1135.407327] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1135.407591] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1b74935-92ce-40cc-99dc-bdd9c888974a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.429720] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1135.429720] env[65788]: value = "task-4663161" [ 1135.429720] env[65788]: _type = "Task" [ 1135.429720] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.441072] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663161, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.612732] env[65788]: DEBUG nova.compute.manager [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Received event network-changed-731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1135.612865] env[65788]: DEBUG nova.compute.manager [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Refreshing instance network info cache due to event network-changed-731c93dc-1d06-4943-bac0-fc800f9f449d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1135.613066] env[65788]: DEBUG oslo_concurrency.lockutils [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Acquiring lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.613214] env[65788]: DEBUG oslo_concurrency.lockutils [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Acquired lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.613365] env[65788]: DEBUG nova.network.neutron [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Refreshing network info cache for port 731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1135.630741] env[65788]: DEBUG oslo_vmware.api [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663154, 'name': ReconfigVM_Task, 'duration_secs': 5.851533} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.630999] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.631220] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Reconfigured VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1135.631666] env[65788]: WARNING neutronclient.v2_0.client [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.631993] env[65788]: WARNING neutronclient.v2_0.client [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.632603] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.632942] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.643569] env[65788]: DEBUG nova.compute.manager [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1135.767209] env[65788]: DEBUG nova.compute.utils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1135.769291] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1135.769565] env[65788]: DEBUG nova.network.neutron [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1135.770011] env[65788]: WARNING neutronclient.v2_0.client [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.770512] env[65788]: WARNING neutronclient.v2_0.client [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1135.771726] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1135.772345] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1135.821676] env[65788]: DEBUG nova.policy [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1135.900830] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520350ca-a4b3-24cc-f179-82c2a81c4079, 'name': SearchDatastore_Task, 'duration_secs': 0.011513} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.902256] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95f208b6-37ff-46ce-bf6c-40f788ff4300 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.910068] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1135.910068] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524270ac-f8e1-1a37-8cf5-43e398e02146" [ 1135.910068] env[65788]: _type = "Task" [ 1135.910068] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.921529] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524270ac-f8e1-1a37-8cf5-43e398e02146, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.942050] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663161, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.117445] env[65788]: WARNING neutronclient.v2_0.client [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.118961] env[65788]: WARNING openstack [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.119541] env[65788]: WARNING openstack [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.144130] env[65788]: DEBUG nova.network.neutron [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Successfully created port: 502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1136.217210] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.283571] env[65788]: WARNING openstack [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.284241] env[65788]: WARNING openstack [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.291437] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1136.352781] env[65788]: WARNING neutronclient.v2_0.client [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.353593] env[65788]: WARNING openstack [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1136.354108] env[65788]: WARNING openstack [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1136.371422] env[65788]: INFO nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating resource usage from migration b175a0c6-bdf7-452a-b15d-6abb54724613 [ 1136.374131] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.374720] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.416329] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance efe1048b-50e9-4add-910a-607a95759c7a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.416449] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1a701758-a056-4948-9069-2a7168f2cc8c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.416570] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9d2f0d5e-793d-4790-ad31-be3d372be4fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.416688] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.416801] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.416913] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 18be7322-3359-49c9-a181-f9228d5c16d7 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417034] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance b9d8a32e-1f0f-405c-9543-9e53046d248e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417151] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 9946cef4-b8c7-41b4-9299-eace1929bf3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417260] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 27ed2344-c42b-46bc-b51a-20821f67cbf0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417369] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 6012182d-8b4e-4379-949f-d3a26211699b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417474] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 50c70e36-a3dc-4e8c-8fe6-74039d1828bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417696] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 84a4e3bf-76cf-4699-8996-230c227cc1f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.417696] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Migration b175a0c6-bdf7-452a-b15d-6abb54724613 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1136.417796] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 5621bbe8-131f-4266-9109-a3be445310a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1136.427480] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524270ac-f8e1-1a37-8cf5-43e398e02146, 'name': SearchDatastore_Task, 'duration_secs': 0.01323} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.427650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.429443] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. {{(pid=65788) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1136.429443] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57625725-16bc-4ff5-93f7-50ae904faf63 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.437725] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1136.437725] env[65788]: value = "task-4663162" [ 1136.437725] env[65788]: _type = "Task" [ 1136.437725] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.442053] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663161, 'name': CreateVM_Task, 'duration_secs': 0.607273} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.446862] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1136.447769] env[65788]: WARNING neutronclient.v2_0.client [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1136.448162] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.448322] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.448640] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1136.449346] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dbcf0d1-75c6-4cae-9841-4a6d2fdc8edb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.455475] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.457419] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1136.457419] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e5721c-e2c1-27b6-4419-66170bad45b2" [ 1136.457419] env[65788]: _type = "Task" [ 1136.457419] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.471175] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e5721c-e2c1-27b6-4419-66170bad45b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.550790] env[65788]: DEBUG nova.network.neutron [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updated VIF entry in instance network info cache for port 731c93dc-1d06-4943-bac0-fc800f9f449d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1136.551300] env[65788]: DEBUG nova.network.neutron [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updating instance_info_cache with network_info: [{"id": "731c93dc-1d06-4943-bac0-fc800f9f449d", "address": "fa:16:3e:bb:7d:4f", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap731c93dc-1d", "ovs_interfaceid": "731c93dc-1d06-4943-bac0-fc800f9f449d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1136.804632] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.806099] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.806443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1136.806648] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.806836] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.809166] env[65788]: INFO nova.compute.manager [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Terminating instance [ 1136.878118] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1136.925035] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance bdd09116-31de-491d-b129-c117f898881e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1136.925035] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1136.925035] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=100GB used_disk=12GB total_vcpus=48 used_vcpus=13 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '13', 'num_vm_active': '11', 'num_task_None': '9', 'num_os_type_None': '13', 'num_proj_ffbe44d63c1d432e97849f15615329e1': '3', 'io_workload': '4', 'num_proj_828bafa0635a42c7be55b1fb367e1734': '1', 'num_proj_676963c25a3741998b10790689242dbb': '1', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '1', 'num_proj_2b70d3c8627449eaa6372ebe3bd90233': '1', 'num_proj_021bf6dd87e340969a57e7e5b33268e7': '1', 'num_proj_009f31c7d5bc4d369a8b96e2aa01117a': '2', 'num_task_rescuing': '1', 'num_proj_e452e2f394f549c793ecfb4f2d5cfa57': '1', 'num_task_resize_prep': '1', 'num_proj_e36859c1eb994614b2a77400f811cf50': '1', 'num_vm_building': '2', 'num_task_spawning': '1', 'num_proj_7aa1dec53aa4436daa0f3223b92f9b79': '1', 'num_task_networking': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1136.926430] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.926983] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.927559] env[65788]: DEBUG nova.network.neutron [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1136.957019] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663162, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.973959] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e5721c-e2c1-27b6-4419-66170bad45b2, 'name': SearchDatastore_Task, 'duration_secs': 0.016714} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.976752] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.976995] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1136.977241] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.977399] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.977609] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1136.978089] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43dced16-3014-4c43-8838-1073a6814a77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.988167] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1136.988378] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1136.991681] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-083283dc-cea8-414c-98dc-b757199e260e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.998167] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1136.998167] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525567f5-b6dc-0897-6273-ba4a4ac5b492" [ 1136.998167] env[65788]: _type = "Task" [ 1136.998167] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.010817] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525567f5-b6dc-0897-6273-ba4a4ac5b492, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.053903] env[65788]: DEBUG oslo_concurrency.lockutils [req-a533e9f6-94fe-423e-bb23-7bcd1769859f req-0e88a703-577a-4867-b612-dd75fb542572 service nova] Releasing lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.161294] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a75933-ff4e-4cde-b4f8-4dd0bbf271c6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.173774] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ee399f-71cd-4ec9-a880-bd6696294416 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.226695] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd73614-9c13-4e0b-9614-0d4b2a8d21e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.238301] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3557ab9b-1047-44f0-a5fa-b8901e669ffb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.262460] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.310816] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1137.314380] env[65788]: DEBUG nova.compute.manager [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1137.314913] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1137.315532] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f5ed40-d4f4-4d89-bd0d-3657dc15ba80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.324166] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.324448] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e747e86-d19c-4723-b34f-6aa2d4feab2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.331787] env[65788]: DEBUG oslo_vmware.api [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1137.331787] env[65788]: value = "task-4663163" [ 1137.331787] env[65788]: _type = "Task" [ 1137.331787] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.340864] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1137.341123] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1137.341283] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1137.341477] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1137.341616] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1137.341761] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1137.341989] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1137.342179] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1137.342351] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1137.342527] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1137.342710] env[65788]: DEBUG nova.virt.hardware [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1137.343983] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff5a212-5099-4773-ae97-1694a0209653 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.349672] env[65788]: DEBUG oslo_vmware.api [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663163, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.355931] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68918a37-844c-4d1d-bbdf-dfcd92b07c00 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.400171] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.433683] env[65788]: WARNING neutronclient.v2_0.client [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.433864] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.434269] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.452776] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564922} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.453020] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. [ 1137.453856] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b265fca-b20f-4ba6-baf7-94424eca98f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.479342] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.482080] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05228cf0-71b9-42c4-8a1f-8d331eb5b149 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.505585] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1137.505585] env[65788]: value = "task-4663164" [ 1137.505585] env[65788]: _type = "Task" [ 1137.505585] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.510401] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525567f5-b6dc-0897-6273-ba4a4ac5b492, 'name': SearchDatastore_Task, 'duration_secs': 0.015146} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.513458] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43ed021e-af49-4220-90bb-f63d4e0f24ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.523859] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1137.523859] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521eadcd-3111-ac0a-120a-e560962cac42" [ 1137.523859] env[65788]: _type = "Task" [ 1137.523859] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.527329] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663164, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.537131] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521eadcd-3111-ac0a-120a-e560962cac42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.620877] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.621428] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.676633] env[65788]: DEBUG nova.compute.manager [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-vif-deleted-a99b6060-25f2-4e56-bc5b-2be77fe833dc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1137.676633] env[65788]: INFO nova.compute.manager [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Neutron deleted interface a99b6060-25f2-4e56-bc5b-2be77fe833dc; detaching it from the instance and deleting it from the info cache [ 1137.676633] env[65788]: DEBUG nova.network.neutron [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1137.768077] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1137.774125] env[65788]: WARNING neutronclient.v2_0.client [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1137.774792] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1137.775231] env[65788]: WARNING openstack [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1137.793538] env[65788]: DEBUG nova.network.neutron [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Successfully updated port: 502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1137.845479] env[65788]: DEBUG oslo_vmware.api [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663163, 'name': PowerOffVM_Task, 'duration_secs': 0.23619} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.845649] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1137.845708] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1137.845960] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e55fd6d1-92de-4d19-ac7e-61f087bf6985 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.873433] env[65788]: INFO nova.network.neutron [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Port a99b6060-25f2-4e56-bc5b-2be77fe833dc from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1137.873820] env[65788]: DEBUG nova.network.neutron [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [{"id": "1426fd1f-0507-498d-9a05-9bce1efd861a", "address": "fa:16:3e:a6:4b:70", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1426fd1f-05", "ovs_interfaceid": "1426fd1f-0507-498d-9a05-9bce1efd861a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1137.918406] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1137.918648] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1137.918832] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleting the datastore file [datastore2] 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.919131] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f6e38be-ff2c-4219-a633-582d51ab00d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.926920] env[65788]: DEBUG oslo_vmware.api [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1137.926920] env[65788]: value = "task-4663166" [ 1137.926920] env[65788]: _type = "Task" [ 1137.926920] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.935353] env[65788]: DEBUG oslo_vmware.api [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.020474] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663164, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.037549] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521eadcd-3111-ac0a-120a-e560962cac42, 'name': SearchDatastore_Task, 'duration_secs': 0.018705} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.037878] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.038218] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 50c70e36-a3dc-4e8c-8fe6-74039d1828bf/50c70e36-a3dc-4e8c-8fe6-74039d1828bf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1138.038505] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40f4a5b5-0543-42f7-acca-9d3ff703902b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.046029] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1138.046029] env[65788]: value = "task-4663167" [ 1138.046029] env[65788]: _type = "Task" [ 1138.046029] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.055884] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663167, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.181525] env[65788]: DEBUG oslo_concurrency.lockutils [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] Acquiring lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.273824] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1138.274168] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.921s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.274494] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.058s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.276206] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.276348] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Cleaning up deleted instances {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11916}} [ 1138.297055] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-84a4e3bf-76cf-4699-8996-230c227cc1f8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.299587] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-84a4e3bf-76cf-4699-8996-230c227cc1f8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.299587] env[65788]: DEBUG nova.network.neutron [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1138.377095] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1138.440596] env[65788]: DEBUG oslo_vmware.api [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170483} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.440902] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.441213] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.441476] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.441756] env[65788]: INFO nova.compute.manager [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1138.442203] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1138.442450] env[65788]: DEBUG nova.compute.manager [-] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1138.442555] env[65788]: DEBUG nova.network.neutron [-] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1138.442861] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.444082] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.444082] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.501566] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1138.526866] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663164, 'name': ReconfigVM_Task, 'duration_secs': 0.831004} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.527470] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.528991] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d623868e-8733-41a5-b983-295d3668e3b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.566950] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-529ebd43-9a26-4414-ab5f-b641c015b136 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.583653] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663167, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.585393] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1138.585393] env[65788]: value = "task-4663168" [ 1138.585393] env[65788]: _type = "Task" [ 1138.585393] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.594833] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663168, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.781378] env[65788]: INFO nova.compute.claims [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1138.798521] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] There are 60 instances to clean {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11925}} [ 1138.798692] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 068bb374-08a4-4f36-a9fc-bb4082f9eceb] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1138.805247] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.805667] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1138.883496] env[65788]: DEBUG nova.network.neutron [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1138.886953] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4daf61b0-5aed-4726-87bf-9dda080964e5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7-a99b6060-25f2-4e56-bc5b-2be77fe833dc" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.870s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.974040] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1138.974539] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.013781] env[65788]: DEBUG nova.compute.manager [req-ee6b5671-1a92-4545-aca7-64372721ca45 req-0f67f966-0497-450a-9075-03a58dd26a34 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Received event network-vif-deleted-1426fd1f-0507-498d-9a05-9bce1efd861a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1139.014454] env[65788]: INFO nova.compute.manager [req-ee6b5671-1a92-4545-aca7-64372721ca45 req-0f67f966-0497-450a-9075-03a58dd26a34 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Neutron deleted interface 1426fd1f-0507-498d-9a05-9bce1efd861a; detaching it from the instance and deleting it from the info cache [ 1139.014643] env[65788]: DEBUG nova.network.neutron [req-ee6b5671-1a92-4545-aca7-64372721ca45 req-0f67f966-0497-450a-9075-03a58dd26a34 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1139.046475] env[65788]: WARNING neutronclient.v2_0.client [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1139.047210] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1139.047651] env[65788]: WARNING openstack [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1139.067501] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663167, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554876} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.068148] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 50c70e36-a3dc-4e8c-8fe6-74039d1828bf/50c70e36-a3dc-4e8c-8fe6-74039d1828bf.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1139.068148] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1139.068466] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b011d55c-693b-4750-8dda-a67a771701bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.076598] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1139.076598] env[65788]: value = "task-4663169" [ 1139.076598] env[65788]: _type = "Task" [ 1139.076598] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.086344] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663169, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.100322] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663168, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.146893] env[65788]: DEBUG nova.network.neutron [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Updating instance_info_cache with network_info: [{"id": "502c664a-0111-47ac-9cc0-f90e3bfa3e99", "address": "fa:16:3e:32:38:60", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap502c664a-01", "ovs_interfaceid": "502c664a-0111-47ac-9cc0-f90e3bfa3e99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1139.303896] env[65788]: INFO nova.compute.resource_tracker [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating resource usage from migration b175a0c6-bdf7-452a-b15d-6abb54724613 [ 1139.307518] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: a022cb81-5ff8-41aa-9307-c99a58416e94] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1139.462990] env[65788]: DEBUG nova.network.neutron [-] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1139.518261] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e03dd56-e11f-4cbb-940b-5dca64ed5601 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.529152] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c829e602-925a-4302-977b-75414173114d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.541796] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6676111-f45b-436d-bb71-541048078fc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.550348] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bd6fe8-7389-4522-824c-cd2d5654c6c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.594853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78119219-a0cb-4a82-93c1-c98f649d94b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.597819] env[65788]: DEBUG nova.compute.manager [req-ee6b5671-1a92-4545-aca7-64372721ca45 req-0f67f966-0497-450a-9075-03a58dd26a34 service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Detach interface failed, port_id=1426fd1f-0507-498d-9a05-9bce1efd861a, reason: Instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1139.612094] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f8233a-dcd1-4fd4-9d08-362df3c9d664 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.617113] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663168, 'name': ReconfigVM_Task, 'duration_secs': 0.55003} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.617374] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663169, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100609} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.617626] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1139.617875] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1139.618521] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a49fcac-4212-432d-906b-5d3c681b1b56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.620386] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f5951b-a34b-4c57-bf35-75bbfbe7ed85 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.630451] env[65788]: DEBUG nova.compute.provider_tree [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.652057] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 50c70e36-a3dc-4e8c-8fe6-74039d1828bf/50c70e36-a3dc-4e8c-8fe6-74039d1828bf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.654781] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-84a4e3bf-76cf-4699-8996-230c227cc1f8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1139.655426] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Instance network_info: |[{"id": "502c664a-0111-47ac-9cc0-f90e3bfa3e99", "address": "fa:16:3e:32:38:60", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap502c664a-01", "ovs_interfaceid": "502c664a-0111-47ac-9cc0-f90e3bfa3e99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1139.655835] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eea1d2a7-469b-42af-abf3-7e770947ca14 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.672476] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1139.672476] env[65788]: value = "task-4663170" [ 1139.672476] env[65788]: _type = "Task" [ 1139.672476] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.672886] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:38:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '502c664a-0111-47ac-9cc0-f90e3bfa3e99', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.681063] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1139.681726] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.686157] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c248648-23ca-4971-bd4e-f63f5b31b331 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.704461] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1139.704461] env[65788]: value = "task-4663171" [ 1139.704461] env[65788]: _type = "Task" [ 1139.704461] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.714570] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663170, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.714819] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.714819] env[65788]: value = "task-4663172" [ 1139.714819] env[65788]: _type = "Task" [ 1139.714819] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.715983] env[65788]: DEBUG nova.compute.manager [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Received event network-vif-plugged-502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1139.716205] env[65788]: DEBUG oslo_concurrency.lockutils [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Acquiring lock "84a4e3bf-76cf-4699-8996-230c227cc1f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.716406] env[65788]: DEBUG oslo_concurrency.lockutils [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.716566] env[65788]: DEBUG oslo_concurrency.lockutils [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1139.716730] env[65788]: DEBUG nova.compute.manager [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] No waiting events found dispatching network-vif-plugged-502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1139.716887] env[65788]: WARNING nova.compute.manager [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Received unexpected event network-vif-plugged-502c664a-0111-47ac-9cc0-f90e3bfa3e99 for instance with vm_state building and task_state spawning. [ 1139.717050] env[65788]: DEBUG nova.compute.manager [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Received event network-changed-502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1139.717205] env[65788]: DEBUG nova.compute.manager [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Refreshing instance network info cache due to event network-changed-502c664a-0111-47ac-9cc0-f90e3bfa3e99. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1139.717381] env[65788]: DEBUG oslo_concurrency.lockutils [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Acquiring lock "refresh_cache-84a4e3bf-76cf-4699-8996-230c227cc1f8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.717511] env[65788]: DEBUG oslo_concurrency.lockutils [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Acquired lock "refresh_cache-84a4e3bf-76cf-4699-8996-230c227cc1f8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1139.717660] env[65788]: DEBUG nova.network.neutron [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Refreshing network info cache for port 502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1139.726805] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663171, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.733723] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663172, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.811191] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2d74b823-6e28-444d-a80b-c91c2d595460] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1139.968281] env[65788]: INFO nova.compute.manager [-] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Took 1.53 seconds to deallocate network for instance. [ 1140.134506] env[65788]: DEBUG nova.scheduler.client.report [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.191577] env[65788]: DEBUG oslo_vmware.api [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663170, 'name': PowerOnVM_Task, 'duration_secs': 0.520579} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.191906] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1140.194772] env[65788]: DEBUG nova.compute.manager [None req-9fd6f32b-f868-440f-b4c2-932fcf4d80c5 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1140.195581] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecac625-c61f-4882-b363-dbdf76ca12dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.213977] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663171, 'name': ReconfigVM_Task, 'duration_secs': 0.443937} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.214296] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 50c70e36-a3dc-4e8c-8fe6-74039d1828bf/50c70e36-a3dc-4e8c-8fe6-74039d1828bf.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1140.214925] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3b555f9-b266-4817-b932-91d62a37c947 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.220802] env[65788]: WARNING neutronclient.v2_0.client [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.221485] env[65788]: WARNING openstack [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.221833] env[65788]: WARNING openstack [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.236158] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1140.236158] env[65788]: value = "task-4663173" [ 1140.236158] env[65788]: _type = "Task" [ 1140.236158] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.243245] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663172, 'name': CreateVM_Task, 'duration_secs': 0.388283} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.243789] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1140.244386] env[65788]: WARNING neutronclient.v2_0.client [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.244747] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.244895] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1140.245447] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1140.249476] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6eb238d-e58a-4d40-839e-33f65a8ee24a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.251702] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663173, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.256473] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1140.256473] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525921a3-a409-8e40-b228-06641d620acf" [ 1140.256473] env[65788]: _type = "Task" [ 1140.256473] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.268275] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525921a3-a409-8e40-b228-06641d620acf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.314585] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 398791b2-9dc0-478a-8ed8-bdfbace0404e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1140.330050] env[65788]: WARNING openstack [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.330452] env[65788]: WARNING openstack [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.397989] env[65788]: WARNING neutronclient.v2_0.client [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1140.398732] env[65788]: WARNING openstack [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1140.399092] env[65788]: WARNING openstack [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1140.478261] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.482334] env[65788]: DEBUG nova.network.neutron [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Updated VIF entry in instance network info cache for port 502c664a-0111-47ac-9cc0-f90e3bfa3e99. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1140.482586] env[65788]: DEBUG nova.network.neutron [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Updating instance_info_cache with network_info: [{"id": "502c664a-0111-47ac-9cc0-f90e3bfa3e99", "address": "fa:16:3e:32:38:60", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap502c664a-01", "ovs_interfaceid": "502c664a-0111-47ac-9cc0-f90e3bfa3e99", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1140.641416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.367s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.641619] env[65788]: INFO nova.compute.manager [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Migrating [ 1140.648864] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.249s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.650457] env[65788]: INFO nova.compute.claims [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.747308] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663173, 'name': Rename_Task, 'duration_secs': 0.171348} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.747521] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1140.747775] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b43d509-1548-4f6d-aa4d-096f95613c12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.755652] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1140.755652] env[65788]: value = "task-4663174" [ 1140.755652] env[65788]: _type = "Task" [ 1140.755652] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.767456] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663174, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.771325] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525921a3-a409-8e40-b228-06641d620acf, 'name': SearchDatastore_Task, 'duration_secs': 0.011708} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.771657] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1140.771971] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1140.772376] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.772554] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1140.772777] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.773059] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f06889c2-b42c-4f56-88e7-33ce84e830ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.783804] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.784024] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1140.784811] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ddde22-9f45-4db4-aa1e-f001fd69d1aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.791194] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1140.791194] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520db816-9fa7-fa81-8abd-560b538b3b22" [ 1140.791194] env[65788]: _type = "Task" [ 1140.791194] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.801188] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520db816-9fa7-fa81-8abd-560b538b3b22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.817816] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 8755c56a-0f5e-4e95-a8b8-c643849fa798] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1140.988812] env[65788]: DEBUG oslo_concurrency.lockutils [req-20a1b48e-e68b-4d45-93b6-c8e119d1c1d9 req-86ac2b13-2327-49f6-aace-3c917c3ee0ac service nova] Releasing lock "refresh_cache-84a4e3bf-76cf-4699-8996-230c227cc1f8" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1141.163561] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.163750] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.163925] env[65788]: DEBUG nova.network.neutron [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1141.268472] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663174, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.302906] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520db816-9fa7-fa81-8abd-560b538b3b22, 'name': SearchDatastore_Task, 'duration_secs': 0.013257} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.303812] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96efba1c-29cc-4c18-ba8a-c17b526261ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.311482] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1141.311482] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218655a-3d4f-ada2-fecb-69a9fb0bb484" [ 1141.311482] env[65788]: _type = "Task" [ 1141.311482] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.320175] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218655a-3d4f-ada2-fecb-69a9fb0bb484, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.320591] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: c5dfbb9b-430d-4e93-b24d-e918d90e123e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1141.491689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.491958] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.526980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "18be7322-3359-49c9-a181-f9228d5c16d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.527334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.527476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.527689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.527887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.530354] env[65788]: INFO nova.compute.manager [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Terminating instance [ 1141.670029] env[65788]: WARNING neutronclient.v2_0.client [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.670498] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.670936] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.769718] env[65788]: DEBUG oslo_vmware.api [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663174, 'name': PowerOnVM_Task, 'duration_secs': 0.843748} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.770158] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1141.770208] env[65788]: INFO nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Took 8.59 seconds to spawn the instance on the hypervisor. [ 1141.770384] env[65788]: DEBUG nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1141.771250] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607cc3f1-78bd-4cae-9c39-e9739906672b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.775119] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.775480] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.823249] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5218655a-3d4f-ada2-fecb-69a9fb0bb484, 'name': SearchDatastore_Task, 'duration_secs': 0.020766} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.826105] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: bc39a402-1f5a-4a67-b09f-a18159cf4abf] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1141.827949] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1141.828240] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 84a4e3bf-76cf-4699-8996-230c227cc1f8/84a4e3bf-76cf-4699-8996-230c227cc1f8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1141.829446] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0ad4dab-599c-46f6-98b1-c078c246927f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.837736] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1141.837736] env[65788]: value = "task-4663175" [ 1141.837736] env[65788]: _type = "Task" [ 1141.837736] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.844264] env[65788]: WARNING neutronclient.v2_0.client [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1141.844892] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1141.845254] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1141.858644] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.923260] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b187b867-9131-4d73-bc69-35a322c75bad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.929971] env[65788]: DEBUG nova.network.neutron [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [{"id": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "address": "fa:16:3e:3d:6c:91", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1870b5-8b", "ovs_interfaceid": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1141.935421] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bbab55-b84e-4d75-aee4-f54327c0cdba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.965310] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be8e8d0-4a3e-43cc-8a7f-115bbc5a7f59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.974264] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa365b5c-8e77-4889-91d7-29abc342139d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.988922] env[65788]: DEBUG nova.compute.provider_tree [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.994956] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1142.034638] env[65788]: DEBUG nova.compute.manager [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1142.034872] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1142.035496] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea69c39f-a585-4e14-97cc-f4fefec37e71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.045308] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1142.045308] env[65788]: value = "task-4663176" [ 1142.045308] env[65788]: _type = "Task" [ 1142.045308] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.054979] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663176, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.289344] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.289814] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.290044] env[65788]: INFO nova.compute.manager [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Shelving [ 1142.302522] env[65788]: INFO nova.compute.manager [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Took 14.38 seconds to build instance. [ 1142.332954] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 31a6981e-89cd-4b83-85cc-36dd163e1f8f] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1142.350948] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663175, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.435196] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1142.492942] env[65788]: DEBUG nova.scheduler.client.report [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.518455] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1142.556190] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663176, 'name': PowerOffVM_Task, 'duration_secs': 0.22338} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.556402] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1142.556665] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1142.556920] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910380', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'name': 'volume-84d81198-187d-46ec-870e-36818081a75e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '18be7322-3359-49c9-a181-f9228d5c16d7', 'attached_at': '2025-11-21T13:54:27.000000', 'detached_at': '', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'serial': '84d81198-187d-46ec-870e-36818081a75e'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1142.557791] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb3aecd-177f-484c-9520-53786c0d15b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.577951] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ccd8d2-1962-4418-99ce-3a0f9addd30d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.586365] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e26093c-6f47-4e5a-911c-c97e3b14f7af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.605397] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a6b54a-2c6a-4382-98f6-02647adfa589 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.622484] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] The volume has not been displaced from its original location: [datastore2] volume-84d81198-187d-46ec-870e-36818081a75e/volume-84d81198-187d-46ec-870e-36818081a75e.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1142.628614] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfiguring VM instance instance-0000005e to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1142.628614] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8335d2f-809d-4d1b-b2a0-be6e46d99b3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.649419] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1142.649419] env[65788]: value = "task-4663177" [ 1142.649419] env[65788]: _type = "Task" [ 1142.649419] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.658528] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663177, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.805705] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3a446eaa-b33c-483a-80a2-b899af4e0903 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.901s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.836958] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 64acef8e-640d-4195-9519-ce31d984ddde] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1142.850492] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551329} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.850751] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 84a4e3bf-76cf-4699-8996-230c227cc1f8/84a4e3bf-76cf-4699-8996-230c227cc1f8.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1142.851017] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1142.851344] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5dda811f-810b-4967-af1a-270e23b79654 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.858948] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1142.858948] env[65788]: value = "task-4663178" [ 1142.858948] env[65788]: _type = "Task" [ 1142.858948] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.869773] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.905644] env[65788]: DEBUG nova.compute.manager [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Received event network-changed-731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1142.905842] env[65788]: DEBUG nova.compute.manager [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Refreshing instance network info cache due to event network-changed-731c93dc-1d06-4943-bac0-fc800f9f449d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1142.906059] env[65788]: DEBUG oslo_concurrency.lockutils [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Acquiring lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.906197] env[65788]: DEBUG oslo_concurrency.lockutils [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Acquired lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.906437] env[65788]: DEBUG nova.network.neutron [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Refreshing network info cache for port 731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1142.997826] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.998984] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1143.001394] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.523s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1143.001674] env[65788]: DEBUG nova.objects.instance [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'resources' on Instance uuid 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1143.160838] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663177, 'name': ReconfigVM_Task, 'duration_secs': 0.21161} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.161182] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Reconfigured VM instance instance-0000005e to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1143.167364] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bdfe836-cd62-4641-a674-c9eaeccd4df0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.187096] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1143.187096] env[65788]: value = "task-4663179" [ 1143.187096] env[65788]: _type = "Task" [ 1143.187096] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.197212] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663179, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.300242] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1143.300647] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27c60f10-6dd2-4829-bc50-1ce87129d82c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.308736] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1143.308736] env[65788]: value = "task-4663180" [ 1143.308736] env[65788]: _type = "Task" [ 1143.308736] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.318015] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.345581] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 51705add-4d88-40bb-b236-2486650c77f4] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1143.371242] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.271771} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.371756] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1143.372657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f50a58e-87fe-4d4e-888b-ef68a0bd3a2e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.396918] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 84a4e3bf-76cf-4699-8996-230c227cc1f8/84a4e3bf-76cf-4699-8996-230c227cc1f8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1143.397564] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edf88a2c-1a34-489b-9c4e-e5bd1e69a92d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.412289] env[65788]: WARNING neutronclient.v2_0.client [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.413119] env[65788]: WARNING openstack [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.413119] env[65788]: WARNING openstack [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.430865] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1143.430865] env[65788]: value = "task-4663181" [ 1143.430865] env[65788]: _type = "Task" [ 1143.430865] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.441432] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663181, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.504879] env[65788]: DEBUG nova.compute.utils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1143.510145] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1143.510145] env[65788]: DEBUG nova.network.neutron [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1143.510400] env[65788]: WARNING neutronclient.v2_0.client [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.510875] env[65788]: WARNING neutronclient.v2_0.client [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.511530] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.512769] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.561253] env[65788]: WARNING openstack [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.561747] env[65788]: WARNING openstack [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.584253] env[65788]: DEBUG nova.policy [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a5c272c62484b1d8c5a3dc9325631ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '94977b75768444138fb914cabddc57c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1143.629103] env[65788]: WARNING neutronclient.v2_0.client [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1143.629785] env[65788]: WARNING openstack [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1143.630221] env[65788]: WARNING openstack [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1143.681450] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.681688] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1143.681898] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.682084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1143.682246] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.686374] env[65788]: INFO nova.compute.manager [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Terminating instance [ 1143.698193] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663179, 'name': ReconfigVM_Task, 'duration_secs': 0.374385} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.701305] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910380', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'name': 'volume-84d81198-187d-46ec-870e-36818081a75e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': '18be7322-3359-49c9-a181-f9228d5c16d7', 'attached_at': '2025-11-21T13:54:27.000000', 'detached_at': '', 'volume_id': '84d81198-187d-46ec-870e-36818081a75e', 'serial': '84d81198-187d-46ec-870e-36818081a75e'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1143.701582] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1143.702947] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357b7778-9a0d-4559-bd22-47437eaf957c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.711146] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1143.711412] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8552aef2-15db-4d70-98b8-bb2d8ddba264 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.718195] env[65788]: DEBUG nova.network.neutron [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updated VIF entry in instance network info cache for port 731c93dc-1d06-4943-bac0-fc800f9f449d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1143.718534] env[65788]: DEBUG nova.network.neutron [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updating instance_info_cache with network_info: [{"id": "731c93dc-1d06-4943-bac0-fc800f9f449d", "address": "fa:16:3e:bb:7d:4f", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap731c93dc-1d", "ovs_interfaceid": "731c93dc-1d06-4943-bac0-fc800f9f449d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1143.778838] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1143.778838] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1143.779136] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore2] 18be7322-3359-49c9-a181-f9228d5c16d7 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.779213] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9968d60-e93b-4f9a-9651-3c01f953fdb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.784771] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc7e86d-2adf-46e2-ae32-1c382469dee2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.788781] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1143.788781] env[65788]: value = "task-4663183" [ 1143.788781] env[65788]: _type = "Task" [ 1143.788781] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.795726] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce93403-346c-44a4-82fb-ef40eab66b4d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.802465] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.835359] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47725b2-1f16-44dd-94cd-e4ca3d3400ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.844829] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663180, 'name': PowerOffVM_Task, 'duration_secs': 0.309833} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.847969] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1143.849521] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fdd55a-caf6-4473-8630-a8921422c2e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.853278] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc144e6c-cd6e-41a6-b162-491b1ad0b40f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.858027] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 6e78042d-ba10-4629-816f-3a13b2e22d4e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1143.888331] env[65788]: DEBUG nova.compute.provider_tree [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.890782] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94de31bf-b713-40e9-ab03-ef1cd930a293 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.922503] env[65788]: DEBUG nova.network.neutron [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Successfully created port: c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1143.942125] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663181, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.950801] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f247f1-6b7d-4dc9-a8a1-de41f22f695e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.969944] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1144.009974] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1144.193326] env[65788]: DEBUG nova.compute.manager [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1144.193553] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1144.194643] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e004dde-02ef-4ae0-8541-cef7cd7d7cb6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.203075] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1144.203347] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b604646e-bedd-4cda-9045-0ee2c0991b29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.211131] env[65788]: DEBUG oslo_vmware.api [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1144.211131] env[65788]: value = "task-4663184" [ 1144.211131] env[65788]: _type = "Task" [ 1144.211131] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.223762] env[65788]: DEBUG oslo_concurrency.lockutils [req-4ac9b887-bedb-495f-8684-837a5e8cba6c req-9141f555-ddce-46b1-8c07-d3a3c5a6b200 service nova] Releasing lock "refresh_cache-50c70e36-a3dc-4e8c-8fe6-74039d1828bf" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.224289] env[65788]: DEBUG oslo_vmware.api [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4663184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.301459] env[65788]: DEBUG oslo_vmware.api [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090694} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.301821] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1144.302086] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1144.302295] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1144.302594] env[65788]: INFO nova.compute.manager [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Took 2.27 seconds to destroy the instance on the hypervisor. [ 1144.302862] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1144.303122] env[65788]: DEBUG nova.compute.manager [-] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1144.303230] env[65788]: DEBUG nova.network.neutron [-] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1144.303486] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.304104] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1144.304382] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1144.344282] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1144.362478] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 902d8a69-0398-4752-a609-0ac5b4b32e27] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1144.394577] env[65788]: DEBUG nova.scheduler.client.report [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1144.404841] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1144.405185] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-56f7ac6b-ca91-4a23-9cfc-0a4af6b7ce29 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.414458] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1144.414458] env[65788]: value = "task-4663185" [ 1144.414458] env[65788]: _type = "Task" [ 1144.414458] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.426701] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663185, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.443241] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663181, 'name': ReconfigVM_Task, 'duration_secs': 0.822123} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.443412] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 84a4e3bf-76cf-4699-8996-230c227cc1f8/84a4e3bf-76cf-4699-8996-230c227cc1f8.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.444469] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7ca6637-8b0c-4332-86f0-6329bdf9fee0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.452087] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1144.452087] env[65788]: value = "task-4663186" [ 1144.452087] env[65788]: _type = "Task" [ 1144.452087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.462640] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663186, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.476430] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1144.476745] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-907c6081-5b35-4011-95e2-8b5b353986a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.484667] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1144.484667] env[65788]: value = "task-4663187" [ 1144.484667] env[65788]: _type = "Task" [ 1144.484667] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.496039] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663187, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.722262] env[65788]: DEBUG oslo_vmware.api [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4663184, 'name': PowerOffVM_Task, 'duration_secs': 0.218759} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.722590] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1144.722814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1144.723142] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcd73c43-0c26-4bf0-b88f-5b7bac38a992 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.798177] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1144.798433] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1144.798610] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Deleting the datastore file [datastore2] 9d2f0d5e-793d-4790-ad31-be3d372be4fa {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1144.798908] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-976ae66d-d97b-448f-b837-1a60570ff2bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.807789] env[65788]: DEBUG oslo_vmware.api [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for the task: (returnval){ [ 1144.807789] env[65788]: value = "task-4663189" [ 1144.807789] env[65788]: _type = "Task" [ 1144.807789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.817854] env[65788]: DEBUG oslo_vmware.api [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4663189, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.865989] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2bcfbe11-51a2-49b4-b482-02ca332d8c38] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1144.901034] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1144.903814] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.385s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1144.905514] env[65788]: INFO nova.compute.claims [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1144.934475] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663185, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.940750] env[65788]: INFO nova.scheduler.client.report [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted allocations for instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 [ 1144.963248] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663186, 'name': Rename_Task, 'duration_secs': 0.175189} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.963865] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1144.963865] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cc1c348-419f-4eaa-af73-c81f9a3abb01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.971307] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1144.971307] env[65788]: value = "task-4663190" [ 1144.971307] env[65788]: _type = "Task" [ 1144.971307] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.980499] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.996568] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663187, 'name': PowerOffVM_Task, 'duration_secs': 0.264345} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.996853] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1144.996992] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1145.022033] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1145.089485] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1145.089812] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1145.089976] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1145.090175] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1145.090351] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1145.090464] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1145.090672] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.090856] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1145.091149] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1145.091434] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1145.091737] env[65788]: DEBUG nova.virt.hardware [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1145.093613] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d900537-7cf1-4c19-a55a-477466139ce3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.106021] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af29558-90fc-4bf3-9611-c6fec88b0c4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.322560] env[65788]: DEBUG oslo_vmware.api [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Task: {'id': task-4663189, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270669} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.322753] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.322940] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1145.323142] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1145.323319] env[65788]: INFO nova.compute.manager [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1145.323587] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1145.323781] env[65788]: DEBUG nova.compute.manager [-] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1145.323873] env[65788]: DEBUG nova.network.neutron [-] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1145.324194] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.324797] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1145.325120] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1145.369618] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 4ffc6bb2-0a25-48b8-9ca5-7ef6c023e7cf] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1145.383649] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1145.428419] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663185, 'name': CreateSnapshot_Task, 'duration_secs': 0.730243} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.428711] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1145.429474] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f70b6c-fb7d-46d2-8a6d-514ae4279795 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.447586] env[65788]: DEBUG oslo_concurrency.lockutils [None req-39752bda-8b3c-42d5-bf8c-781cd5038db5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.643s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.448476] env[65788]: DEBUG oslo_concurrency.lockutils [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] Acquired lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.449444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67986e35-89f3-4689-ac4b-e290745ae040 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.459759] env[65788]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1145.459759] env[65788]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=65788) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1145.460067] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67c9ee7c-483f-4c42-a577-b464d9ca764f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.470305] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a79910-76d0-4b2d-9f79-95a475d60a25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.491530] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.495192] env[65788]: DEBUG nova.network.neutron [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Successfully updated port: c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1145.512510] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1145.512742] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1145.512895] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1145.513083] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1145.513233] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1145.513376] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1145.513573] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.513726] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1145.513884] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1145.514051] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1145.514231] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1145.521571] env[65788]: ERROR root [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-910400' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 479, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-910400' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-910400' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-910400'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-910400' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-910400' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-910400'}\n"]: nova.exception.InstanceNotFound: Instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 could not be found. [ 1145.521873] env[65788]: DEBUG oslo_concurrency.lockutils [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] Releasing lock "1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.521990] env[65788]: DEBUG nova.compute.manager [req-e7caec60-6894-4a48-8a6a-85e4d93b55a7 req-f34aba8b-416b-496c-8856-be54dc0b77cb service nova] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Detach interface failed, port_id=a99b6060-25f2-4e56-bc5b-2be77fe833dc, reason: Instance 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1145.522435] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-620687b4-1f26-4bcb-9ed1-2c51ebcbee53 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.539297] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1145.539297] env[65788]: value = "task-4663191" [ 1145.539297] env[65788]: _type = "Task" [ 1145.539297] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.549671] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663191, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.771780] env[65788]: DEBUG nova.network.neutron [-] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1145.810757] env[65788]: DEBUG nova.compute.manager [req-0f6e67fe-c7bf-40dd-b326-ea468be0e453 req-3c1963ed-de31-47fa-b424-e0c5a2f5ef57 service nova] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Received event network-vif-deleted-ab03f5bb-a9ba-42df-bbe4-d8a3b6acd401 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1145.830486] env[65788]: DEBUG nova.compute.manager [req-8908c637-93d0-4df0-b2b8-8035314c6a0e req-9ec64adf-bbfa-42a5-a2d5-4a71b9406703 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-vif-plugged-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1145.830578] env[65788]: DEBUG oslo_concurrency.lockutils [req-8908c637-93d0-4df0-b2b8-8035314c6a0e req-9ec64adf-bbfa-42a5-a2d5-4a71b9406703 service nova] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1145.830739] env[65788]: DEBUG oslo_concurrency.lockutils [req-8908c637-93d0-4df0-b2b8-8035314c6a0e req-9ec64adf-bbfa-42a5-a2d5-4a71b9406703 service nova] Lock "bdd09116-31de-491d-b129-c117f898881e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1145.830901] env[65788]: DEBUG oslo_concurrency.lockutils [req-8908c637-93d0-4df0-b2b8-8035314c6a0e req-9ec64adf-bbfa-42a5-a2d5-4a71b9406703 service nova] Lock "bdd09116-31de-491d-b129-c117f898881e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1145.831102] env[65788]: DEBUG nova.compute.manager [req-8908c637-93d0-4df0-b2b8-8035314c6a0e req-9ec64adf-bbfa-42a5-a2d5-4a71b9406703 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] No waiting events found dispatching network-vif-plugged-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1145.831299] env[65788]: WARNING nova.compute.manager [req-8908c637-93d0-4df0-b2b8-8035314c6a0e req-9ec64adf-bbfa-42a5-a2d5-4a71b9406703 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received unexpected event network-vif-plugged-c6074f79-c6f2-4186-b2c3-809f045737ed for instance with vm_state building and task_state spawning. [ 1145.873157] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: af9120a6-2cea-46da-ba06-6036b9beda78] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1145.952374] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1145.956831] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b3189ce1-ed3e-4b2f-a9ec-a3102fe39b03 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.967090] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1145.967090] env[65788]: value = "task-4663192" [ 1145.967090] env[65788]: _type = "Task" [ 1145.967090] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.979538] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663192, 'name': CloneVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.999436] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.999744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1146.000192] env[65788]: DEBUG nova.network.neutron [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1146.001425] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.053415] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663191, 'name': ReconfigVM_Task, 'duration_secs': 0.326157} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.053770] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.188680] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca310da8-b2df-49d1-9c00-de52c2c073cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.197949] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0bedfc1-001b-498d-a0be-9b98a49bca12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.232348] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276fbff2-559d-4cfe-ae33-8a4a534b36b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.243511] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e369a4b-17a0-4fb4-82df-725ce5603619 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.260354] env[65788]: DEBUG nova.compute.provider_tree [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.275167] env[65788]: INFO nova.compute.manager [-] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Took 1.97 seconds to deallocate network for instance. [ 1146.359032] env[65788]: DEBUG nova.network.neutron [-] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1146.377725] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 22379ac4-fe2f-4fcc-aa2b-c6b15b33e7dc] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1146.480635] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663192, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.494078] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.506426] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.506980] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.553670] env[65788]: DEBUG nova.network.neutron [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1146.563677] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1146.564070] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1146.564271] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1146.564509] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1146.564703] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1146.564923] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1146.565077] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.565305] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1146.565422] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1146.565582] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1146.565751] env[65788]: DEBUG nova.virt.hardware [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1146.571381] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1146.571747] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76130a3b-0c79-43a2-afe7-839418402d1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.587028] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.587486] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.600876] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1146.600876] env[65788]: value = "task-4663193" [ 1146.600876] env[65788]: _type = "Task" [ 1146.600876] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.609876] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663193, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.662268] env[65788]: WARNING neutronclient.v2_0.client [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1146.662961] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1146.663327] env[65788]: WARNING openstack [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1146.764436] env[65788]: DEBUG nova.scheduler.client.report [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1146.826929] env[65788]: INFO nova.compute.manager [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Took 0.55 seconds to detach 1 volumes for instance. [ 1146.831121] env[65788]: DEBUG nova.compute.manager [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Deleting volume: 84d81198-187d-46ec-870e-36818081a75e {{(pid=65788) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3320}} [ 1146.861708] env[65788]: INFO nova.compute.manager [-] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Took 1.54 seconds to deallocate network for instance. [ 1146.873588] env[65788]: DEBUG nova.network.neutron [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1146.886843] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: b220df77-cd3a-4aeb-87ad-68f56995ec3b] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1146.983952] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663192, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.994758] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.112509] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663193, 'name': ReconfigVM_Task, 'duration_secs': 0.18564} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.112833] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1147.113764] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2986f17-7d96-4e40-a40d-bd2f1a9b0847 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.139944] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0/5621bbe8-131f-4266-9109-a3be445310a0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.140472] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25fb6317-ae45-42bd-b688-33863541997a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.162603] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1147.162603] env[65788]: value = "task-4663195" [ 1147.162603] env[65788]: _type = "Task" [ 1147.162603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.172159] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.269339] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1147.269892] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1147.369995] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1147.370465] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.370817] env[65788]: DEBUG nova.objects.instance [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lazy-loading 'resources' on Instance uuid 9d2f0d5e-793d-4790-ad31-be3d372be4fa {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.377433] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1147.377578] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance network_info: |[{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1147.378981] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:d9:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a0a76279-3c11-4bef-b124-2a2ee13fa377', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6074f79-c6f2-4186-b2c3-809f045737ed', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1147.386385] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1147.387251] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1147.387567] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ce62800-cc4f-481b-aea1-659572075fc5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.403652] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1147.405373] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: a79bbb7e-7b78-4900-9b83-bfce94bb41bd] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1147.413951] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.413951] env[65788]: value = "task-4663196" [ 1147.413951] env[65788]: _type = "Task" [ 1147.413951] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.425882] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663196, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.482062] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663192, 'name': CloneVM_Task, 'duration_secs': 1.387942} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.482639] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Created linked-clone VM from snapshot [ 1147.483616] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b2060a-dab8-4fb2-bc88-f75d6e4bb98e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.496537] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Uploading image aa3a5ddd-40cd-4c98-8384-9c40c7390261 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1147.503811] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.530333] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1147.530333] env[65788]: value = "vm-910424" [ 1147.530333] env[65788]: _type = "VirtualMachine" [ 1147.530333] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1147.530657] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-47167154-90ba-4d5f-88fa-8478fc6375a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.542100] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease: (returnval){ [ 1147.542100] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52196b32-79ea-2e4d-3f76-683311d0832e" [ 1147.542100] env[65788]: _type = "HttpNfcLease" [ 1147.542100] env[65788]: } obtained for exporting VM: (result){ [ 1147.542100] env[65788]: value = "vm-910424" [ 1147.542100] env[65788]: _type = "VirtualMachine" [ 1147.542100] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1147.542614] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the lease: (returnval){ [ 1147.542614] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52196b32-79ea-2e4d-3f76-683311d0832e" [ 1147.542614] env[65788]: _type = "HttpNfcLease" [ 1147.542614] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1147.551438] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1147.551438] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52196b32-79ea-2e4d-3f76-683311d0832e" [ 1147.551438] env[65788]: _type = "HttpNfcLease" [ 1147.551438] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1147.640865] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1147.641361] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1147.673651] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663195, 'name': ReconfigVM_Task, 'duration_secs': 0.411428} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.673955] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0/5621bbe8-131f-4266-9109-a3be445310a0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1147.674306] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1147.775331] env[65788]: DEBUG nova.compute.utils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1147.776750] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1147.776943] env[65788]: DEBUG nova.network.neutron [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1147.777281] env[65788]: WARNING neutronclient.v2_0.client [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.777586] env[65788]: WARNING neutronclient.v2_0.client [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.778236] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1147.778596] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1147.838094] env[65788]: DEBUG nova.policy [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b220b7a1fb504d3c9b2cfd5a4b940dac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e452e2f394f549c793ecfb4f2d5cfa57', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1147.842150] env[65788]: DEBUG nova.compute.manager [req-bbdb53cc-ba87-4c3f-bdce-d153ddb0e669 req-87aa778f-abe1-4db5-85ae-399dc43ddc17 service nova] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Received event network-vif-deleted-b382ad0a-1833-409d-8d27-6726195bc924 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1147.909315] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: c07e6ee2-6ba0-4cb9-b2b6-5d1c001961eb] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1147.928019] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663196, 'name': CreateVM_Task, 'duration_secs': 0.366781} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.928019] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1147.928163] env[65788]: WARNING neutronclient.v2_0.client [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1147.928513] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.930053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1147.930053] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1147.930053] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f123fd37-6a08-48fa-bb53-686ed8df4349 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.939457] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1147.939457] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ac0ae-9679-f2ee-1e94-ccd787525cfd" [ 1147.939457] env[65788]: _type = "Task" [ 1147.939457] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.949679] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ac0ae-9679-f2ee-1e94-ccd787525cfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.956945] env[65788]: DEBUG nova.compute.manager [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1147.957342] env[65788]: DEBUG nova.compute.manager [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing instance network info cache due to event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1147.957743] env[65788]: DEBUG oslo_concurrency.lockutils [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.958080] env[65788]: DEBUG oslo_concurrency.lockutils [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1147.958343] env[65788]: DEBUG nova.network.neutron [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1147.997298] env[65788]: DEBUG oslo_vmware.api [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663190, 'name': PowerOnVM_Task, 'duration_secs': 2.673179} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.000192] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1148.000436] env[65788]: INFO nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Took 10.69 seconds to spawn the instance on the hypervisor. [ 1148.000615] env[65788]: DEBUG nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1148.001767] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ece5277-80c9-457e-88b1-e22393e2c04d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.052487] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1148.052487] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52196b32-79ea-2e4d-3f76-683311d0832e" [ 1148.052487] env[65788]: _type = "HttpNfcLease" [ 1148.052487] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1148.052487] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1148.052487] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52196b32-79ea-2e4d-3f76-683311d0832e" [ 1148.052487] env[65788]: _type = "HttpNfcLease" [ 1148.052487] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1148.052879] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc982e7-7ccd-4384-9e1f-02f61587e1f6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.068023] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523f9fb4-4369-bd5a-4b22-274ed9beece4/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1148.068023] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523f9fb4-4369-bd5a-4b22-274ed9beece4/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1148.146860] env[65788]: DEBUG nova.compute.utils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1148.161619] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d31201f9-ab0a-4c66-b214-2e230c3d4386 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.184191] env[65788]: DEBUG nova.network.neutron [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Successfully created port: 458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1148.187316] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e701c3f-ae72-4f0f-83a5-e4fcd41579a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.216659] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1148.216659] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "f8c57215-ade6-424b-be8c-075998a764af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.220818] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bcf701a-ca71-467d-b1bc-6cfa3a0c7515 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.224213] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c380be68-151c-4451-9d9c-15edd99c9a59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.248843] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba43d8e3-5de6-4bd6-b603-f73a073321d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.252982] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1148.291979] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1148.295140] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfbfc2b-5eff-4959-9f4a-0b9e485cc341 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.304342] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a244c8a4-3e25-4569-823e-a48514a61904 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.320420] env[65788]: DEBUG nova.compute.provider_tree [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.412409] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: ec7f7a46-8dd2-4b75-866f-20e73907f1cd] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1148.455455] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529ac0ae-9679-f2ee-1e94-ccd787525cfd, 'name': SearchDatastore_Task, 'duration_secs': 0.010365} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.456346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1148.456619] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1148.456889] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.457096] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1148.457323] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1148.457904] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3956202-7588-44d9-a012-e4daecf8e4c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.462364] env[65788]: WARNING neutronclient.v2_0.client [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.463462] env[65788]: WARNING openstack [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.463937] env[65788]: WARNING openstack [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.477062] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1148.477399] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1148.478475] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f66099c-7aff-4afe-9b06-0d8debedfbd4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.486712] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1148.486712] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a5dbf-c10c-5086-1428-bce2c7cbff95" [ 1148.486712] env[65788]: _type = "Task" [ 1148.486712] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.498248] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a5dbf-c10c-5086-1428-bce2c7cbff95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.528596] env[65788]: INFO nova.compute.manager [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Took 19.72 seconds to build instance. [ 1148.640509] env[65788]: WARNING openstack [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.640995] env[65788]: WARNING openstack [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.650042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.728032] env[65788]: WARNING neutronclient.v2_0.client [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.728781] env[65788]: WARNING openstack [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1148.729173] env[65788]: WARNING openstack [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1148.738070] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1148.764492] env[65788]: WARNING neutronclient.v2_0.client [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1148.819732] env[65788]: DEBUG nova.network.neutron [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Port ec1870b5-8b77-4ad8-9779-fdcd725dd800 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1148.824027] env[65788]: DEBUG nova.scheduler.client.report [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.878539] env[65788]: DEBUG nova.network.neutron [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updated VIF entry in instance network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1148.879167] env[65788]: DEBUG nova.network.neutron [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1148.915360] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: e0f7c849-315c-4247-a840-d388d48746b9] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1148.999232] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524a5dbf-c10c-5086-1428-bce2c7cbff95, 'name': SearchDatastore_Task, 'duration_secs': 0.015481} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.001488] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8df836fa-1c86-484a-aca3-0c15cf0f4d3e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.009247] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1149.009247] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5291bc03-3220-9e84-2c14-bf43637e1f63" [ 1149.009247] env[65788]: _type = "Task" [ 1149.009247] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.021024] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5291bc03-3220-9e84-2c14-bf43637e1f63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.030444] env[65788]: DEBUG oslo_concurrency.lockutils [None req-409e6972-7d1a-439b-bf04-caad4c9fc6ea tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.231s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.265495] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.304704] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1149.330038] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.332807] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.929s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.333093] env[65788]: DEBUG nova.objects.instance [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'resources' on Instance uuid 18be7322-3359-49c9-a181-f9228d5c16d7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.342070] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1149.344079] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1149.344800] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1149.345208] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1149.345705] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1149.345874] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1149.346104] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1149.346273] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1149.346438] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1149.346838] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1149.347279] env[65788]: DEBUG nova.virt.hardware [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1149.348436] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18d8c78-d032-44d8-be2d-ca669f3003e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.364499] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca67e65-f5c8-470e-b9a6-530c02978b86 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.370255] env[65788]: INFO nova.scheduler.client.report [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Deleted allocations for instance 9d2f0d5e-793d-4790-ad31-be3d372be4fa [ 1149.387158] env[65788]: DEBUG oslo_concurrency.lockutils [req-e31b20b6-0c47-4adb-b7ea-2ac48e08c481 req-00b1e56b-4493-4481-a30d-97c731b0d63e service nova] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1149.418880] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 4dcf8335-701d-4386-9de4-f14f5d375d1f] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1149.484944] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "84a4e3bf-76cf-4699-8996-230c227cc1f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.485404] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.485754] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "84a4e3bf-76cf-4699-8996-230c227cc1f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.486084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.486383] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.489988] env[65788]: INFO nova.compute.manager [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Terminating instance [ 1149.521785] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5291bc03-3220-9e84-2c14-bf43637e1f63, 'name': SearchDatastore_Task, 'duration_secs': 0.015936} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.522188] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1149.522586] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1149.522908] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80b09146-7fb4-4b23-8e20-f2f0de20951d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.532032] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1149.532032] env[65788]: value = "task-4663198" [ 1149.532032] env[65788]: _type = "Task" [ 1149.532032] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.542457] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663198, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.785214] env[65788]: DEBUG nova.network.neutron [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Successfully updated port: 458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1149.851262] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.851612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.851854] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.865274] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1149.865529] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1149.865896] env[65788]: INFO nova.compute.manager [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Attaching volume ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b to /dev/sdb [ 1149.893923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc8e3f2e-71d3-484b-b628-757912ab0cb7 tempest-ServersTestFqdnHostnames-296510809 tempest-ServersTestFqdnHostnames-296510809-project-member] Lock "9d2f0d5e-793d-4790-ad31-be3d372be4fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.212s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1149.906320] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c246fbfe-5e4d-4dd2-b97d-f8d9a72cbbab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.918465] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3007126c-d1aa-4567-8313-fd8c710ea6bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.922338] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 6a69d645-f89d-43b6-8bdc-b12ce0d88ca8] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1149.940283] env[65788]: DEBUG nova.virt.block_device [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updating existing volume attachment record: 6e23fe35-f2fe-41b9-a726-14a94a6c62ab {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1149.997643] env[65788]: DEBUG nova.compute.manager [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1149.997731] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1149.999242] env[65788]: DEBUG nova.compute.manager [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Received event network-vif-plugged-458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1150.000029] env[65788]: DEBUG oslo_concurrency.lockutils [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Acquiring lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1150.000029] env[65788]: DEBUG oslo_concurrency.lockutils [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1150.000283] env[65788]: DEBUG oslo_concurrency.lockutils [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1150.000670] env[65788]: DEBUG nova.compute.manager [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] No waiting events found dispatching network-vif-plugged-458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1150.000866] env[65788]: WARNING nova.compute.manager [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Received unexpected event network-vif-plugged-458887b5-5ecd-41a3-9bb5-4ee7926f28ad for instance with vm_state building and task_state spawning. [ 1150.000942] env[65788]: DEBUG nova.compute.manager [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Received event network-changed-458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1150.001110] env[65788]: DEBUG nova.compute.manager [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Refreshing instance network info cache due to event network-changed-458887b5-5ecd-41a3-9bb5-4ee7926f28ad. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1150.001432] env[65788]: DEBUG oslo_concurrency.lockutils [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Acquiring lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.001432] env[65788]: DEBUG oslo_concurrency.lockutils [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Acquired lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.001622] env[65788]: DEBUG nova.network.neutron [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Refreshing network info cache for port 458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1150.009150] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618865ba-9ef0-4c3b-a9f6-362ae59c5862 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.020221] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1150.021263] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebec6922-f8c9-4af3-9458-9ebbaf445adc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.029962] env[65788]: DEBUG oslo_vmware.api [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1150.029962] env[65788]: value = "task-4663199" [ 1150.029962] env[65788]: _type = "Task" [ 1150.029962] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.051666] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663198, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.051863] env[65788]: DEBUG oslo_vmware.api [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.103964] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca88cf3-ee3e-463d-841f-79cdc8665fdf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.116685] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6043b72-8e9a-4237-a11f-b4862a49e31f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.163810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70eaaeb-25b0-4847-a3bc-f4545ae82ba2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.173956] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9fe191-686c-4918-bc58-07d62463ce74 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.192320] env[65788]: DEBUG nova.compute.provider_tree [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.288250] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.425587] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2cb53fe5-aa58-479e-9090-0d8509cf164e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1150.514070] env[65788]: WARNING neutronclient.v2_0.client [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.514991] env[65788]: WARNING openstack [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1150.515518] env[65788]: WARNING openstack [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1150.542405] env[65788]: DEBUG oslo_vmware.api [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663199, 'name': PowerOffVM_Task, 'duration_secs': 0.263207} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.542405] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1150.542667] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1150.542965] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d07e7a5f-4a2b-48a9-90b4-aa526c734422 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.549020] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663198, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.728137} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.549751] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1150.550036] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1150.550553] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c4a0815-8e56-4999-ba31-6e1683c64981 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.555472] env[65788]: DEBUG nova.network.neutron [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1150.562664] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1150.562664] env[65788]: value = "task-4663202" [ 1150.562664] env[65788]: _type = "Task" [ 1150.562664] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.573865] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.624551] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1150.625301] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1150.625775] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore2] 84a4e3bf-76cf-4699-8996-230c227cc1f8 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1150.626254] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d18c460-2f70-4eda-99f7-51b6344d49eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.634082] env[65788]: DEBUG oslo_vmware.api [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1150.634082] env[65788]: value = "task-4663203" [ 1150.634082] env[65788]: _type = "Task" [ 1150.634082] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.645894] env[65788]: DEBUG oslo_vmware.api [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.697616] env[65788]: DEBUG nova.scheduler.client.report [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1150.761412] env[65788]: DEBUG nova.network.neutron [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1150.860505] env[65788]: WARNING neutronclient.v2_0.client [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1150.902118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.902118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1150.902118] env[65788]: DEBUG nova.network.neutron [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1150.930790] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 7f637326-9388-4d73-a1c7-3e4611ac46c5] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1151.074019] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.119588} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.074019] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1151.074423] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f595cff-91c3-4809-b66d-cc0fdbfab50c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.098908] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1151.099273] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff20dd67-bc02-4713-8843-29e648e408ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.119713] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1151.119713] env[65788]: value = "task-4663204" [ 1151.119713] env[65788]: _type = "Task" [ 1151.119713] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.129530] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663204, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.146372] env[65788]: DEBUG oslo_vmware.api [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280603} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.146783] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1151.147069] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1151.147379] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1151.147631] env[65788]: INFO nova.compute.manager [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1151.147947] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1151.148325] env[65788]: DEBUG nova.compute.manager [-] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1151.148668] env[65788]: DEBUG nova.network.neutron [-] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1151.148930] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.149939] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.150040] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.195090] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.207124] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.874s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.210181] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.945s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.211846] env[65788]: INFO nova.compute.claims [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1151.235018] env[65788]: INFO nova.scheduler.client.report [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted allocations for instance 18be7322-3359-49c9-a181-f9228d5c16d7 [ 1151.265996] env[65788]: DEBUG oslo_concurrency.lockutils [req-01a63786-aba5-4fad-922e-baa773d07459 req-e7619011-a4b4-4ab8-8bae-e5f5b30e74e8 service nova] Releasing lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.266558] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.266807] env[65788]: DEBUG nova.network.neutron [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1151.406839] env[65788]: WARNING neutronclient.v2_0.client [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.407760] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.408212] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.430592] env[65788]: DEBUG nova.compute.manager [req-883b70ad-cb6c-4165-995e-846390bf8f35 req-80308427-36bc-4625-9da0-2b056068a4e8 service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Received event network-vif-deleted-502c664a-0111-47ac-9cc0-f90e3bfa3e99 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1151.430883] env[65788]: INFO nova.compute.manager [req-883b70ad-cb6c-4165-995e-846390bf8f35 req-80308427-36bc-4625-9da0-2b056068a4e8 service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Neutron deleted interface 502c664a-0111-47ac-9cc0-f90e3bfa3e99; detaching it from the instance and deleting it from the info cache [ 1151.431177] env[65788]: DEBUG nova.network.neutron [req-883b70ad-cb6c-4165-995e-846390bf8f35 req-80308427-36bc-4625-9da0-2b056068a4e8 service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1151.434086] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 374564df-70fa-4b89-8f38-e559245e5ebf] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1151.574245] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.574773] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.632418] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663204, 'name': ReconfigVM_Task, 'duration_secs': 0.439428} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.632664] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Reconfigured VM instance instance-0000006f to attach disk [datastore2] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1151.633359] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0f3f99c-f28c-4e2e-bdb1-9661bb7087e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.642285] env[65788]: WARNING neutronclient.v2_0.client [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.643301] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.643442] env[65788]: WARNING openstack [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.653305] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1151.653305] env[65788]: value = "task-4663205" [ 1151.653305] env[65788]: _type = "Task" [ 1151.653305] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.664210] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663205, 'name': Rename_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.740705] env[65788]: DEBUG nova.network.neutron [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [{"id": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "address": "fa:16:3e:3d:6c:91", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1870b5-8b", "ovs_interfaceid": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1151.749045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b328a9c6-5a0b-448e-b226-a27aa46f6247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "18be7322-3359-49c9-a181-f9228d5c16d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.220s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.770073] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.772049] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.824932] env[65788]: DEBUG nova.network.neutron [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1151.877111] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.878761] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1151.902856] env[65788]: DEBUG nova.network.neutron [-] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1151.935868] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f12ff512-b8be-4b90-b7e3-4e676cef2c70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.938311] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: b4f157a7-350c-4fff-8509-44426714846c] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1151.951393] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f89884-41a7-41a7-a39b-b36d63b29dfe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.993859] env[65788]: DEBUG nova.compute.manager [req-883b70ad-cb6c-4165-995e-846390bf8f35 req-80308427-36bc-4625-9da0-2b056068a4e8 service nova] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Detach interface failed, port_id=502c664a-0111-47ac-9cc0-f90e3bfa3e99, reason: Instance 84a4e3bf-76cf-4699-8996-230c227cc1f8 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1151.995427] env[65788]: WARNING neutronclient.v2_0.client [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1151.996018] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1151.996572] env[65788]: WARNING openstack [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1152.107103] env[65788]: DEBUG nova.network.neutron [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Updating instance_info_cache with network_info: [{"id": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "address": "fa:16:3e:9e:1f:ca", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap458887b5-5e", "ovs_interfaceid": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1152.164686] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663205, 'name': Rename_Task, 'duration_secs': 0.225006} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.165066] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1152.165449] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc3e5d4d-8cf5-45af-90d1-44847c3b8798 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.173878] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1152.173878] env[65788]: value = "task-4663206" [ 1152.173878] env[65788]: _type = "Task" [ 1152.173878] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.183709] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663206, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.246026] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.406420] env[65788]: INFO nova.compute.manager [-] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Took 1.26 seconds to deallocate network for instance. [ 1152.444179] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 9aeb45e5-ebd5-417b-aa9a-4d6914fca1ca] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1152.449211] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09380631-c683-49f5-a655-4940606f51b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.458822] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d546bd-6e90-4dea-b99d-51670de24245 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.495344] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56353cf2-9f66-4519-bdfd-f526cde807dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.504674] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a0fa3e-aaed-4d35-91ef-f62e14a9094e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.521821] env[65788]: DEBUG nova.compute.provider_tree [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.610450] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1152.611575] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Instance network_info: |[{"id": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "address": "fa:16:3e:9e:1f:ca", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap458887b5-5e", "ovs_interfaceid": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1152.611575] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:1f:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56398cc0-e39f-410f-8036-8c2a6870e26f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '458887b5-5ecd-41a3-9bb5-4ee7926f28ad', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1152.619316] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1152.619579] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1152.620241] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5700c3a5-5a1b-4161-8a8a-f13ba50468bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.641823] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1152.641823] env[65788]: value = "task-4663208" [ 1152.641823] env[65788]: _type = "Task" [ 1152.641823] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.651429] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663208, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.687838] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.751847] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "1a701758-a056-4948-9069-2a7168f2cc8c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.752105] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "1a701758-a056-4948-9069-2a7168f2cc8c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.752431] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "1a701758-a056-4948-9069-2a7168f2cc8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.752703] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "1a701758-a056-4948-9069-2a7168f2cc8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1152.752917] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "1a701758-a056-4948-9069-2a7168f2cc8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1152.759571] env[65788]: INFO nova.compute.manager [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Terminating instance [ 1152.778105] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae533b22-3116-41ce-a356-fecc22be8092 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.800727] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fdbd6d-4242-4a6d-a8da-7c869caa6855 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.809352] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1152.916829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1152.949248] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 7c88e3ef-82d2-46ac-a350-999de8091c01] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1153.026331] env[65788]: DEBUG nova.scheduler.client.report [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1153.158301] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663208, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.187974] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663206, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.266339] env[65788]: DEBUG nova.compute.manager [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1153.266724] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1153.267957] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6b511b-d214-413f-a7ed-a3441da8bc30 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.277485] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1153.277819] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eef0d130-2910-4cd7-b6d3-1b0c3f12e5e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.287383] env[65788]: DEBUG oslo_vmware.api [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1153.287383] env[65788]: value = "task-4663209" [ 1153.287383] env[65788]: _type = "Task" [ 1153.287383] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.298976] env[65788]: DEBUG oslo_vmware.api [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.316657] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1153.317443] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b92e252-6a6d-4fa5-bd60-523b7a633de8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.327326] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1153.327326] env[65788]: value = "task-4663210" [ 1153.327326] env[65788]: _type = "Task" [ 1153.327326] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.339743] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.453811] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5e27dbf0-44d2-46a4-87b2-c209afb9559e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1153.533623] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.534630] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1153.540306] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.622s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.540306] env[65788]: DEBUG nova.objects.instance [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid 84a4e3bf-76cf-4699-8996-230c227cc1f8 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.659307] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663208, 'name': CreateVM_Task, 'duration_secs': 0.675124} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.659547] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1153.660330] env[65788]: WARNING neutronclient.v2_0.client [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1153.660941] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.661241] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.661762] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1153.662199] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54a252cc-a16f-4b50-9eee-46569767b37a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.670217] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1153.670217] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cd7b71-d8f8-7fc0-4b51-97ef785372ef" [ 1153.670217] env[65788]: _type = "Task" [ 1153.670217] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.611958] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d8103991-fd1f-4a57-81ce-1a47dc4defe0] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1154.618065] env[65788]: DEBUG nova.compute.utils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1154.620696] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cd7b71-d8f8-7fc0-4b51-97ef785372ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.620976] env[65788]: WARNING oslo_vmware.common.loopingcall [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] task run outlasted interval by 0.450442 sec [ 1154.623383] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1154.623383] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910426', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'name': 'volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'serial': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1154.635585] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1154.635799] env[65788]: DEBUG nova.network.neutron [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1154.636133] env[65788]: WARNING neutronclient.v2_0.client [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.636468] env[65788]: WARNING neutronclient.v2_0.client [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1154.637443] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1154.637443] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1154.647614] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fd6705-065f-4523-add9-ba413055265f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.648458] env[65788]: DEBUG oslo_vmware.api [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663206, 'name': PowerOnVM_Task, 'duration_secs': 1.437786} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.652756] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1154.652991] env[65788]: INFO nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Took 9.63 seconds to spawn the instance on the hypervisor. [ 1154.653189] env[65788]: DEBUG nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1154.657641] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f44654a-51af-4cbc-97d4-b3f5e3a671cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.677571] env[65788]: DEBUG oslo_vmware.api [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663210, 'name': PowerOnVM_Task, 'duration_secs': 0.9426} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.678023] env[65788]: DEBUG oslo_vmware.api [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663209, 'name': PowerOffVM_Task, 'duration_secs': 0.363941} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.681202] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1154.681412] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-3f27d162-d1cc-4521-8463-576aa5395b97 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance '5621bbe8-131f-4266-9109-a3be445310a0' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1154.689032] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d995b0-737c-4ede-9d2d-5481add15eb5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.689475] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1154.689605] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1154.696575] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b08bb55-66d3-4037-af39-2f555c176087 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.701597] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cd7b71-d8f8-7fc0-4b51-97ef785372ef, 'name': SearchDatastore_Task, 'duration_secs': 0.022844} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.705185] env[65788]: DEBUG nova.policy [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1154.707650] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1154.707927] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1154.708118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.708258] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.708444] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1154.711954] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92c2a0e0-3ddf-4fbe-aaed-abd59b28bc61 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.736047] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b/volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1154.736541] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de533e40-4413-4695-99e1-066c34f706f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.758975] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1154.759234] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1154.760926] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14da300-d7ef-47ca-a42b-1cf0d0c53aa7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.766754] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1154.766754] env[65788]: value = "task-4663212" [ 1154.766754] env[65788]: _type = "Task" [ 1154.766754] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.768654] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1154.768654] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e6a81e-63da-e19e-334f-975ea71a3c7e" [ 1154.768654] env[65788]: _type = "Task" [ 1154.768654] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.786015] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e6a81e-63da-e19e-334f-975ea71a3c7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.789677] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663212, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.799918] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1154.800129] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1154.801282] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore2] 1a701758-a056-4948-9069-2a7168f2cc8c {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.801282] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-052b5bf1-a83e-4e7e-9b4f-678128b53f38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.810091] env[65788]: DEBUG oslo_vmware.api [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1154.810091] env[65788]: value = "task-4663213" [ 1154.810091] env[65788]: _type = "Task" [ 1154.810091] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.823513] env[65788]: DEBUG oslo_vmware.api [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.924374] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d812749c-0af7-46fb-a01e-f0ef4c0ec90f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.934601] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-258ec4ea-0aae-40ba-8e7f-47663f3fe5e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.968027] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f5f7b2-b8d4-4a06-a476-61dfd7a4fa22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.976945] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5428abf2-f45c-4e82-ad6b-e2a1a4d21f19 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.993125] env[65788]: DEBUG nova.compute.provider_tree [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.108735] env[65788]: DEBUG nova.network.neutron [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Successfully created port: fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1155.121249] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1155.125534] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 34a7e0cc-ec4a-4d1a-b3ce-2cc00744f468] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1155.215204] env[65788]: INFO nova.compute.manager [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Took 17.83 seconds to build instance. [ 1155.290610] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e6a81e-63da-e19e-334f-975ea71a3c7e, 'name': SearchDatastore_Task, 'duration_secs': 0.025639} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.290820] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663212, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.291709] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-794a2ece-2585-40d5-80cf-b4ac12cf392c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.298042] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1155.298042] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52265d96-ee49-f885-b79a-680205b1e1e9" [ 1155.298042] env[65788]: _type = "Task" [ 1155.298042] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.307547] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52265d96-ee49-f885-b79a-680205b1e1e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.320161] env[65788]: DEBUG oslo_vmware.api [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321736} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.320423] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.320601] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1155.320771] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1155.320933] env[65788]: INFO nova.compute.manager [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Took 2.05 seconds to destroy the instance on the hypervisor. [ 1155.321294] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1155.321523] env[65788]: DEBUG nova.compute.manager [-] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1155.321623] env[65788]: DEBUG nova.network.neutron [-] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1155.322239] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.322565] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1155.322833] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1155.496576] env[65788]: DEBUG nova.scheduler.client.report [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1155.633038] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 8d7a33e6-08fb-481e-8fac-fdf13b19aa21] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1155.707517] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1155.717553] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a0630a65-7c4c-4284-9178-53e706d353a0 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.343s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.783664] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663212, 'name': ReconfigVM_Task, 'duration_secs': 0.810649} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.783664] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b/volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1155.787522] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6e3f498-e27e-4c5d-8400-7c60030be924 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.811629] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52265d96-ee49-f885-b79a-680205b1e1e9, 'name': SearchDatastore_Task, 'duration_secs': 0.01921} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.813116] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.813398] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/f59d609a-e5bb-4202-ac05-d486bbf9e05b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1155.813735] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1155.813735] env[65788]: value = "task-4663214" [ 1155.813735] env[65788]: _type = "Task" [ 1155.813735] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.813935] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-acee8ff0-2a90-4a5e-ba8f-77d90ca063a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.826356] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663214, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.827113] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1155.827113] env[65788]: value = "task-4663215" [ 1155.827113] env[65788]: _type = "Task" [ 1155.827113] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.836328] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663215, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.002395] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.463s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.050133] env[65788]: INFO nova.scheduler.client.report [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance 84a4e3bf-76cf-4699-8996-230c227cc1f8 [ 1156.138035] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1156.139438] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5ea56ab6-7b0c-4376-b40a-45a8bf27bd7a] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1156.186611] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1156.187795] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1156.187795] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1156.187795] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1156.187795] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1156.187795] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1156.187795] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1156.187999] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1156.188163] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1156.188339] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1156.188513] env[65788]: DEBUG nova.virt.hardware [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1156.189931] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1814288-3686-4622-b7fb-26e6c931a968 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.207288] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443237de-15aa-4147-93c5-1c80e0f2ca06 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.327803] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.340832] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663215, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.454022] env[65788]: DEBUG nova.compute.manager [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1156.454022] env[65788]: DEBUG nova.compute.manager [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing instance network info cache due to event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1156.454022] env[65788]: DEBUG oslo_concurrency.lockutils [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.454022] env[65788]: DEBUG oslo_concurrency.lockutils [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1156.454022] env[65788]: DEBUG nova.network.neutron [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1156.458059] env[65788]: DEBUG nova.network.neutron [-] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1156.562777] env[65788]: DEBUG oslo_concurrency.lockutils [None req-87bccd3d-d8fa-4ce2-b33e-9c4d3d675a2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "84a4e3bf-76cf-4699-8996-230c227cc1f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.077s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.645034] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d2d614f8-15d9-47d6-9e66-8e1f8fcc18da] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1156.682202] env[65788]: DEBUG nova.compute.manager [req-165da1e0-78d4-496a-8123-efd01cfa687f req-21475e72-8f42-40af-afa5-759a8f5cc5cc service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-vif-plugged-fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1156.682535] env[65788]: DEBUG oslo_concurrency.lockutils [req-165da1e0-78d4-496a-8123-efd01cfa687f req-21475e72-8f42-40af-afa5-759a8f5cc5cc service nova] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.682808] env[65788]: DEBUG oslo_concurrency.lockutils [req-165da1e0-78d4-496a-8123-efd01cfa687f req-21475e72-8f42-40af-afa5-759a8f5cc5cc service nova] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.683057] env[65788]: DEBUG oslo_concurrency.lockutils [req-165da1e0-78d4-496a-8123-efd01cfa687f req-21475e72-8f42-40af-afa5-759a8f5cc5cc service nova] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.683705] env[65788]: DEBUG nova.compute.manager [req-165da1e0-78d4-496a-8123-efd01cfa687f req-21475e72-8f42-40af-afa5-759a8f5cc5cc service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] No waiting events found dispatching network-vif-plugged-fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1156.683705] env[65788]: WARNING nova.compute.manager [req-165da1e0-78d4-496a-8123-efd01cfa687f req-21475e72-8f42-40af-afa5-759a8f5cc5cc service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received unexpected event network-vif-plugged-fda83fac-56b6-4ab0-824a-7af792b7c8c7 for instance with vm_state building and task_state spawning. [ 1156.812542] env[65788]: DEBUG nova.network.neutron [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Successfully updated port: fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1156.832104] env[65788]: DEBUG oslo_vmware.api [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663214, 'name': ReconfigVM_Task, 'duration_secs': 0.897249} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.837435] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910426', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'name': 'volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'serial': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1156.846468] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663215, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.957848] env[65788]: WARNING neutronclient.v2_0.client [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1156.958618] env[65788]: WARNING openstack [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1156.958781] env[65788]: WARNING openstack [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1156.972222] env[65788]: INFO nova.compute.manager [-] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Took 1.65 seconds to deallocate network for instance. [ 1157.079977] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.080218] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.105761] env[65788]: WARNING openstack [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.106173] env[65788]: WARNING openstack [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.149037] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 7fd30bbe-5797-4d0f-8d01-5c8a953f2ada] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1157.199938] env[65788]: WARNING neutronclient.v2_0.client [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1157.200663] env[65788]: WARNING openstack [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.201112] env[65788]: WARNING openstack [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.320538] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.320538] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1157.320538] env[65788]: DEBUG nova.network.neutron [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1157.328352] env[65788]: DEBUG nova.network.neutron [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updated VIF entry in instance network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1157.328721] env[65788]: DEBUG nova.network.neutron [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1157.347539] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663215, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.195617} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.347798] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/f59d609a-e5bb-4202-ac05-d486bbf9e05b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1157.348155] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1157.348296] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14de639c-4bf2-46a5-97d1-cbbdca4f391e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.357831] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1157.357831] env[65788]: value = "task-4663216" [ 1157.357831] env[65788]: _type = "Task" [ 1157.357831] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.370048] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.479049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.479383] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.479652] env[65788]: DEBUG nova.objects.instance [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'resources' on Instance uuid 1a701758-a056-4948-9069-2a7168f2cc8c {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.533319] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.533557] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.533730] env[65788]: DEBUG nova.compute.manager [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Going to confirm migration 8 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1157.583220] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1157.652642] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d042a386-0d16-4e2d-a8e7-7641c07bcb35] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1157.814618] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.814999] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.823790] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.824295] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.835272] env[65788]: DEBUG oslo_concurrency.lockutils [req-29c1e089-9bee-486c-8b85-6b61bdbecbca req-add4af71-b0b3-4831-b926-9468a1a37457 service nova] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1157.866226] env[65788]: DEBUG nova.network.neutron [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1157.871629] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.158367} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.871884] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1157.872735] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698c5772-249a-4431-a320-20a45ac6a65b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.900529] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/f59d609a-e5bb-4202-ac05-d486bbf9e05b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1157.902160] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1157.902541] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1157.909538] env[65788]: DEBUG nova.objects.instance [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.911139] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceaa2a05-1e43-4260-bc71-f86903715800 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.933552] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1157.933552] env[65788]: value = "task-4663217" [ 1157.933552] env[65788]: _type = "Task" [ 1157.933552] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.946774] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.016702] env[65788]: WARNING neutronclient.v2_0.client [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.017777] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1158.018241] env[65788]: WARNING openstack [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1158.042968] env[65788]: WARNING neutronclient.v2_0.client [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.102271] env[65788]: WARNING neutronclient.v2_0.client [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1158.102601] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.102741] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquired lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.106020] env[65788]: DEBUG nova.network.neutron [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1158.106020] env[65788]: DEBUG nova.objects.instance [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'info_cache' on Instance uuid 5621bbe8-131f-4266-9109-a3be445310a0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.113173] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.150805] env[65788]: DEBUG nova.network.neutron [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1158.156185] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 3a06a833-5aaa-4b5d-88b3-8a1d469580af] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1158.289074] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25db4d2f-3427-4432-9b1a-6f8ebdc8cf69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.301614] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73175dbf-71f2-48d5-a5c8-d2c3bdc5290b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.334464] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1158.338470] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d611520a-681a-4b03-9074-26d945f3497f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.347809] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f08c1f0-3941-4933-962c-313b7c31dffc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.365679] env[65788]: DEBUG nova.compute.provider_tree [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.429220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-a262e78c-06ef-4d11-a011-89336f7fe7f8 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.563s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1158.446251] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663217, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.546824] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523f9fb4-4369-bd5a-4b22-274ed9beece4/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1158.547794] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d08e76-4315-4e49-b1ab-5707a9687dc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.556029] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523f9fb4-4369-bd5a-4b22-274ed9beece4/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1158.556029] env[65788]: ERROR oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523f9fb4-4369-bd5a-4b22-274ed9beece4/disk-0.vmdk due to incomplete transfer. [ 1158.556362] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-be1ba602-35b8-4e31-89e4-4fb8d8be5ff0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.566320] env[65788]: DEBUG nova.compute.manager [req-76e8ebc0-5312-4032-94fa-a949ab4f807f req-c51876fe-9705-4c20-acf4-de83a8a8dc8a service nova] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Received event network-vif-deleted-9f372d9b-f11d-40fc-b795-bae66f4a1800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1158.568177] env[65788]: DEBUG oslo_vmware.rw_handles [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523f9fb4-4369-bd5a-4b22-274ed9beece4/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1158.568367] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Uploaded image aa3a5ddd-40cd-4c98-8384-9c40c7390261 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1158.571304] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1158.571907] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6e18454f-659c-436a-8528-7eac2066d776 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.582036] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1158.582036] env[65788]: value = "task-4663218" [ 1158.582036] env[65788]: _type = "Task" [ 1158.582036] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.590915] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663218, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.653560] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1158.653968] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Instance network_info: |[{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1158.654513] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:90:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fda83fac-56b6-4ab0-824a-7af792b7c8c7', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1158.662139] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1158.662392] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8c57215-ade6-424b-be8c-075998a764af] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1158.662790] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: a35ced42-4317-49b4-b4cc-4ed7e2c85c64] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1158.664655] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69e5a6e3-030f-402a-abe2-15346f5b8e81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.687256] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1158.687256] env[65788]: value = "task-4663219" [ 1158.687256] env[65788]: _type = "Task" [ 1158.687256] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.700239] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663219, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.791327] env[65788]: DEBUG nova.compute.manager [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-changed-fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1158.791549] env[65788]: DEBUG nova.compute.manager [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing instance network info cache due to event network-changed-fda83fac-56b6-4ab0-824a-7af792b7c8c7. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1158.791755] env[65788]: DEBUG oslo_concurrency.lockutils [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.791902] env[65788]: DEBUG oslo_concurrency.lockutils [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.792067] env[65788]: DEBUG nova.network.neutron [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing network info cache for port fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1158.865095] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.867722] env[65788]: DEBUG nova.scheduler.client.report [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1158.946468] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663217, 'name': ReconfigVM_Task, 'duration_secs': 0.712621} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.946625] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Reconfigured VM instance instance-00000070 to attach disk [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/f59d609a-e5bb-4202-ac05-d486bbf9e05b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.947390] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-501ccd56-7b7b-4b8e-ab89-dfa48efff6c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.955981] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1158.955981] env[65788]: value = "task-4663220" [ 1158.955981] env[65788]: _type = "Task" [ 1158.955981] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.966206] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663220, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.092625] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663218, 'name': Destroy_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.111118] env[65788]: WARNING neutronclient.v2_0.client [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.111118] env[65788]: WARNING openstack [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.111486] env[65788]: WARNING openstack [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.180435] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 37471e59-1809-4df3-8c40-20921d04d18e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1159.199065] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663219, 'name': CreateVM_Task} progress is 25%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.238642] env[65788]: WARNING openstack [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.239071] env[65788]: WARNING openstack [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.296372] env[65788]: WARNING neutronclient.v2_0.client [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.296372] env[65788]: WARNING openstack [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.297559] env[65788]: WARNING openstack [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.326123] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.326403] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.329077] env[65788]: WARNING neutronclient.v2_0.client [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.329606] env[65788]: WARNING openstack [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.329934] env[65788]: WARNING openstack [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.372762] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.893s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.375260] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.262s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.378857] env[65788]: INFO nova.compute.claims [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.397419] env[65788]: INFO nova.scheduler.client.report [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted allocations for instance 1a701758-a056-4948-9069-2a7168f2cc8c [ 1159.407731] env[65788]: WARNING openstack [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.408263] env[65788]: WARNING openstack [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.440820] env[65788]: DEBUG nova.network.neutron [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [{"id": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "address": "fa:16:3e:3d:6c:91", "network": {"id": "31e92f9f-b71e-46a7-9f7b-a917b955ec88", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-455553166-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e36859c1eb994614b2a77400f811cf50", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1d468f87-964a-4fb6-bab3-b83f6f2646b5", "external-id": "nsx-vlan-transportzone-131", "segmentation_id": 131, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec1870b5-8b", "ovs_interfaceid": "ec1870b5-8b77-4ad8-9779-fdcd725dd800", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.466828] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663220, 'name': Rename_Task, 'duration_secs': 0.318902} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.467212] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1159.467382] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e98e5804-7229-4ebe-b0f8-c397c2482b6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.478411] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1159.478411] env[65788]: value = "task-4663221" [ 1159.478411] env[65788]: _type = "Task" [ 1159.478411] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.479626] env[65788]: WARNING neutronclient.v2_0.client [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.480249] env[65788]: WARNING openstack [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1159.480598] env[65788]: WARNING openstack [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1159.501077] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663221, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.577024] env[65788]: DEBUG nova.network.neutron [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updated VIF entry in instance network info cache for port fda83fac-56b6-4ab0-824a-7af792b7c8c7. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1159.577168] env[65788]: DEBUG nova.network.neutron [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1159.593635] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663218, 'name': Destroy_Task, 'duration_secs': 0.894018} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.593800] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Destroyed the VM [ 1159.593926] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1159.594235] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8e0457ce-b689-4163-bf66-2e1ac61bedc5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.602619] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1159.602619] env[65788]: value = "task-4663222" [ 1159.602619] env[65788]: _type = "Task" [ 1159.602619] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.611390] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663222, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.684079] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f160e4a5-1e91-495e-800e-116ec435d8e1] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1159.699638] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663219, 'name': CreateVM_Task, 'duration_secs': 0.672597} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.699837] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8c57215-ade6-424b-be8c-075998a764af] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1159.700770] env[65788]: WARNING neutronclient.v2_0.client [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1159.701322] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.701477] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1159.701794] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1159.702126] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7c5bdb8-503d-4cbe-a89e-aff9bd04f740 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.707710] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1159.707710] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526fbc55-260d-9e5c-3dbb-650bbd3f8f67" [ 1159.707710] env[65788]: _type = "Task" [ 1159.707710] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.717172] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526fbc55-260d-9e5c-3dbb-650bbd3f8f67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.840571] env[65788]: DEBUG nova.compute.utils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1159.905598] env[65788]: DEBUG oslo_concurrency.lockutils [None req-714b14cc-e92e-4dde-8f51-820e539bdc41 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "1a701758-a056-4948-9069-2a7168f2cc8c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.153s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.947330] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Releasing lock "refresh_cache-5621bbe8-131f-4266-9109-a3be445310a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.947632] env[65788]: DEBUG nova.objects.instance [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lazy-loading 'migration_context' on Instance uuid 5621bbe8-131f-4266-9109-a3be445310a0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.000462] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663221, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.079747] env[65788]: DEBUG oslo_concurrency.lockutils [req-89ffde15-5928-4e03-9803-cae7721e0f9a req-14b04f1c-8c56-44a6-9ff8-2a4468b41900 service nova] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1160.113549] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663222, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.188497] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: bbf6a242-9873-44b9-8938-2f8df11a1018] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1160.220113] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526fbc55-260d-9e5c-3dbb-650bbd3f8f67, 'name': SearchDatastore_Task, 'duration_secs': 0.043074} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.220710] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1160.220945] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.221206] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.221381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1160.221565] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.222141] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc896e28-8c14-4f69-8668-a50f7a46a974 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.232982] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.233262] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1160.233997] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cda235f-0615-4de7-9950-e27c3295f2dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.243089] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1160.243089] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c50c89-9e4a-de5b-c57d-8ad5a7c5a6cf" [ 1160.243089] env[65788]: _type = "Task" [ 1160.243089] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.252352] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c50c89-9e4a-de5b-c57d-8ad5a7c5a6cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.343607] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.017s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.450828] env[65788]: DEBUG nova.objects.base [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Object Instance<5621bbe8-131f-4266-9109-a3be445310a0> lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1160.451811] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f200acc5-08ca-4345-80f5-4e245c1b9742 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.477454] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9703d094-9e26-48a0-a6b5-8493acf3da9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.483497] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1160.483497] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52afbaca-1a82-0ef4-dc1e-17dc2a95fdc4" [ 1160.483497] env[65788]: _type = "Task" [ 1160.483497] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.493319] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52afbaca-1a82-0ef4-dc1e-17dc2a95fdc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.503900] env[65788]: DEBUG oslo_vmware.api [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663221, 'name': PowerOnVM_Task, 'duration_secs': 1.020459} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.504216] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1160.504430] env[65788]: INFO nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Took 11.20 seconds to spawn the instance on the hypervisor. [ 1160.504610] env[65788]: DEBUG nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1160.505702] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0277ea5-69bf-450e-8ecc-c8ff13446d00 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.604041] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b1be56-420e-4fd0-ab6d-f82009667b80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.618459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3439f035-ebc6-44ea-ab26-7db1548520ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.621548] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663222, 'name': RemoveSnapshot_Task, 'duration_secs': 1.005657} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.621808] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1160.622108] env[65788]: DEBUG nova.compute.manager [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1160.623472] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088759c3-dc38-49e0-8623-42eefe9e6195 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.652022] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f277be-75d8-490a-9ef4-ea8c61bc08dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.664495] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9dd1e1f-220e-4927-80b7-9b9d7f4bca1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.680601] env[65788]: DEBUG nova.compute.provider_tree [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.692876] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 93e1ade8-4c57-4db9-9da3-388e3eea91bd] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1160.755376] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c50c89-9e4a-de5b-c57d-8ad5a7c5a6cf, 'name': SearchDatastore_Task, 'duration_secs': 0.011413} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.756150] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75cfb582-71b6-402c-9fcb-f2991f97b22c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.761973] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1160.761973] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527ecdd5-fc6a-7b9f-cd59-82c0b8b93a33" [ 1160.761973] env[65788]: _type = "Task" [ 1160.761973] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.770348] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527ecdd5-fc6a-7b9f-cd59-82c0b8b93a33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.994983] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52afbaca-1a82-0ef4-dc1e-17dc2a95fdc4, 'name': SearchDatastore_Task, 'duration_secs': 0.009499} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.995355] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.023148] env[65788]: INFO nova.compute.manager [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Took 18.52 seconds to build instance. [ 1161.163146] env[65788]: INFO nova.compute.manager [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Shelve offloading [ 1161.185538] env[65788]: DEBUG nova.scheduler.client.report [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1161.196483] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: caceb296-5da2-4b4d-b955-5d7a238ff939] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1161.273821] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527ecdd5-fc6a-7b9f-cd59-82c0b8b93a33, 'name': SearchDatastore_Task, 'duration_secs': 0.028509} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.274094] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1161.274391] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f8c57215-ade6-424b-be8c-075998a764af/f8c57215-ade6-424b-be8c-075998a764af.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1161.274664] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-44ad0b10-4f7f-4f3d-84ab-7643a8c51fc5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.282627] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1161.282627] env[65788]: value = "task-4663223" [ 1161.282627] env[65788]: _type = "Task" [ 1161.282627] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.293227] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663223, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.407855] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.408208] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.408492] env[65788]: INFO nova.compute.manager [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Attaching volume 814933a5-cdb9-4d95-b486-fd9349ad7a21 to /dev/sdc [ 1161.446347] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36031302-926e-492f-b9f6-247c82b26fa5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.455214] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24b7bb1-9d2b-4423-81f1-0bc19fcf1704 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.472092] env[65788]: DEBUG nova.virt.block_device [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updating existing volume attachment record: 76da2e72-48be-4395-afed-2c645c7cbe54 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1161.525875] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5fc804c5-8ee0-4629-aa01-4e2261f03130 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.034s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.668066] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.668456] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f954110-af5d-4165-8a6a-de7df31cfc36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.678093] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1161.678093] env[65788]: value = "task-4663226" [ 1161.678093] env[65788]: _type = "Task" [ 1161.678093] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.691124] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1161.691821] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1161.694980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.830s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.696603] env[65788]: INFO nova.compute.claims [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1161.699536] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1161.699755] env[65788]: DEBUG nova.compute.manager [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1161.700588] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 47ff3955-5d9b-4d90-b8b1-276a0ce58a21] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1161.702940] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29eaa30a-d7ab-4745-aff4-c6fe2a3e7410 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.711684] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.711862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1161.712100] env[65788]: DEBUG nova.network.neutron [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1161.793433] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663223, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.950114] env[65788]: INFO nova.compute.manager [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Rescuing [ 1161.951011] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.951011] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1161.951011] env[65788]: DEBUG nova.network.neutron [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1162.204085] env[65788]: DEBUG nova.compute.utils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1162.205878] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1162.206780] env[65788]: DEBUG nova.network.neutron [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1162.206780] env[65788]: WARNING neutronclient.v2_0.client [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.207220] env[65788]: WARNING neutronclient.v2_0.client [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.208110] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.208686] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.219711] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1162.222443] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2c24bc30-d413-4714-81c2-b657cafe94bd] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1162.225273] env[65788]: WARNING neutronclient.v2_0.client [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.226356] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.227076] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.274201] env[65788]: DEBUG nova.policy [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98de130f8f5d4ea18c7ebfc8d017fa68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fef7c297452947d7979e7972d6887130', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1162.294294] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663223, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.601552} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.294575] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f8c57215-ade6-424b-be8c-075998a764af/f8c57215-ade6-424b-be8c-075998a764af.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1162.294788] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.295068] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca426d94-fb50-43d7-985d-05b937f5dfd9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.302840] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1162.302840] env[65788]: value = "task-4663227" [ 1162.302840] env[65788]: _type = "Task" [ 1162.302840] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.318943] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663227, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.400358] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.400491] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.455557] env[65788]: WARNING neutronclient.v2_0.client [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.456063] env[65788]: WARNING openstack [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.456513] env[65788]: WARNING openstack [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.475659] env[65788]: WARNING neutronclient.v2_0.client [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.476363] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.476826] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.594812] env[65788]: DEBUG nova.network.neutron [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Successfully created port: 98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1162.609652] env[65788]: WARNING openstack [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.609964] env[65788]: WARNING openstack [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.620127] env[65788]: DEBUG nova.network.neutron [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1162.661080] env[65788]: WARNING neutronclient.v2_0.client [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1162.661765] env[65788]: WARNING openstack [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1162.662809] env[65788]: WARNING openstack [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1162.744819] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: a813d102-1b77-4214-8eab-2cd66e002912] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1162.809538] env[65788]: DEBUG nova.network.neutron [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Updating instance_info_cache with network_info: [{"id": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "address": "fa:16:3e:9e:1f:ca", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap458887b5-5e", "ovs_interfaceid": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1162.822196] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663227, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077605} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.822484] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1162.823523] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d06c99-8b43-4fac-83df-591276e7985a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.851620] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] f8c57215-ade6-424b-be8c-075998a764af/f8c57215-ade6-424b-be8c-075998a764af.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.854768] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4b1b612-8063-49f1-93f7-1b8ea7103450 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.877325] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1162.877325] env[65788]: value = "task-4663228" [ 1162.877325] env[65788]: _type = "Task" [ 1162.877325] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.889598] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663228, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.954774] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029c128f-4e62-4774-a5d6-7fba2fd9ff6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.963588] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf9208c-df4c-402c-8541-71d0d879d7e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.997445] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa1ab78-5ff3-4124-b318-bde4ff525964 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.008998] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7afda23-b6b7-4f01-9d15-d64eb090f123 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.027617] env[65788]: DEBUG nova.compute.provider_tree [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.122486] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.123071] env[65788]: WARNING neutronclient.v2_0.client [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.123867] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1163.124337] env[65788]: WARNING openstack [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1163.130865] env[65788]: WARNING neutronclient.v2_0.client [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1163.249436] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1163.258497] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2f32bb1d-11e2-4469-a1e0-5fbc1baab3d9] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1163.280407] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1163.280778] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1163.281065] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1163.281402] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1163.281672] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1163.282073] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1163.282337] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1163.282604] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1163.282896] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1163.283181] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1163.283486] env[65788]: DEBUG nova.virt.hardware [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1163.285041] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6800fc4-8196-4cf6-9901-ab7e995ae46a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.299209] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4c8e16-94c5-4049-8848-c0da24beab5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.323095] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.388269] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663228, 'name': ReconfigVM_Task, 'duration_secs': 0.330018} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.388593] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfigured VM instance instance-00000071 to attach disk [datastore2] f8c57215-ade6-424b-be8c-075998a764af/f8c57215-ade6-424b-be8c-075998a764af.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.389277] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81e45ff9-def8-49d1-a86f-ae0e9b53c078 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.397733] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1163.397733] env[65788]: value = "task-4663229" [ 1163.397733] env[65788]: _type = "Task" [ 1163.397733] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.408190] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663229, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.532532] env[65788]: DEBUG nova.scheduler.client.report [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1163.587673] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1163.589340] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67f4864-60ec-4650-ad75-1955158c19b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.599156] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1163.599756] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16e7ea37-a93e-4a1e-8574-83d4b8db6c31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.707580] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1163.707814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1163.707997] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleting the datastore file [datastore2] b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1163.708350] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dab04287-773e-4486-8656-224e31def126 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.716365] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1163.716365] env[65788]: value = "task-4663231" [ 1163.716365] env[65788]: _type = "Task" [ 1163.716365] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.727221] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.760335] env[65788]: DEBUG nova.compute.manager [req-0da9c056-93f4-4c65-b4a0-0523347e4546 req-f25fd087-2c91-4b24-84be-88a3952e33f5 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-vif-unplugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1163.760643] env[65788]: DEBUG oslo_concurrency.lockutils [req-0da9c056-93f4-4c65-b4a0-0523347e4546 req-f25fd087-2c91-4b24-84be-88a3952e33f5 service nova] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1163.760910] env[65788]: DEBUG oslo_concurrency.lockutils [req-0da9c056-93f4-4c65-b4a0-0523347e4546 req-f25fd087-2c91-4b24-84be-88a3952e33f5 service nova] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1163.761241] env[65788]: DEBUG oslo_concurrency.lockutils [req-0da9c056-93f4-4c65-b4a0-0523347e4546 req-f25fd087-2c91-4b24-84be-88a3952e33f5 service nova] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1163.761470] env[65788]: DEBUG nova.compute.manager [req-0da9c056-93f4-4c65-b4a0-0523347e4546 req-f25fd087-2c91-4b24-84be-88a3952e33f5 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] No waiting events found dispatching network-vif-unplugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1163.761652] env[65788]: WARNING nova.compute.manager [req-0da9c056-93f4-4c65-b4a0-0523347e4546 req-f25fd087-2c91-4b24-84be-88a3952e33f5 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received unexpected event network-vif-unplugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc for instance with vm_state shelved and task_state shelving_offloading. [ 1163.762531] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f75000e6-f3bf-4951-a541-3aa185339054] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1163.908998] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663229, 'name': Rename_Task, 'duration_secs': 0.324662} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.909319] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1163.909569] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d6a696d-3d01-437d-b032-2b2738f70024 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.917437] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1163.917437] env[65788]: value = "task-4663233" [ 1163.917437] env[65788]: _type = "Task" [ 1163.917437] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.927864] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.022898] env[65788]: DEBUG nova.compute.manager [req-5156819d-1f41-421d-9d39-54643816bd62 req-5c309127-5619-4962-89d1-9de56c331ae0 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Received event network-vif-plugged-98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1164.023180] env[65788]: DEBUG oslo_concurrency.lockutils [req-5156819d-1f41-421d-9d39-54643816bd62 req-5c309127-5619-4962-89d1-9de56c331ae0 service nova] Acquiring lock "22cd2208-ebb5-401a-a06d-3dd6cd591869-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.023391] env[65788]: DEBUG oslo_concurrency.lockutils [req-5156819d-1f41-421d-9d39-54643816bd62 req-5c309127-5619-4962-89d1-9de56c331ae0 service nova] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.023557] env[65788]: DEBUG oslo_concurrency.lockutils [req-5156819d-1f41-421d-9d39-54643816bd62 req-5c309127-5619-4962-89d1-9de56c331ae0 service nova] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.023727] env[65788]: DEBUG nova.compute.manager [req-5156819d-1f41-421d-9d39-54643816bd62 req-5c309127-5619-4962-89d1-9de56c331ae0 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] No waiting events found dispatching network-vif-plugged-98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1164.023886] env[65788]: WARNING nova.compute.manager [req-5156819d-1f41-421d-9d39-54643816bd62 req-5c309127-5619-4962-89d1-9de56c331ae0 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Received unexpected event network-vif-plugged-98f4edb9-9923-4aa2-9ee0-e4db751e1691 for instance with vm_state building and task_state spawning. [ 1164.038237] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.038782] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1164.041685] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 3.046s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.122573] env[65788]: DEBUG nova.network.neutron [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Successfully updated port: 98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1164.228278] env[65788]: DEBUG oslo_vmware.api [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218771} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.228598] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.228785] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1164.228965] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1164.254107] env[65788]: INFO nova.scheduler.client.report [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted allocations for instance b9d8a32e-1f0f-405c-9543-9e53046d248e [ 1164.265582] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d78d36e6-a542-4ba8-9e29-b8a7b28c559e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1164.428585] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663233, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.548190] env[65788]: DEBUG nova.compute.utils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1164.552198] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1164.552463] env[65788]: DEBUG nova.network.neutron [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1164.552745] env[65788]: WARNING neutronclient.v2_0.client [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.553055] env[65788]: WARNING neutronclient.v2_0.client [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1164.553672] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1164.554047] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1164.562139] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1164.609097] env[65788]: DEBUG nova.policy [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1164.625396] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "refresh_cache-22cd2208-ebb5-401a-a06d-3dd6cd591869" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.625660] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquired lock "refresh_cache-22cd2208-ebb5-401a-a06d-3dd6cd591869" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.625855] env[65788]: DEBUG nova.network.neutron [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1164.758816] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.759227] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.759433] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.759664] env[65788]: INFO nova.compute.manager [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Rebooting instance [ 1164.768829] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 263cd2fb-7eeb-4395-b1e9-b829940ebd35] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1164.781909] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c8023a-cfcb-46ac-ab90-f6b861e6610c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.791350] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f5edad-291f-47fc-a574-a0a013c28bbc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.827125] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67a7a9a-4ff1-469a-9024-c700f9367d65 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.836591] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59096144-3499-4adc-b845-340902106456 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.852011] env[65788]: DEBUG nova.compute.provider_tree [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.858856] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1164.859014] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a81a6425-ea15-4118-aceb-9e088ee09ea3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.868058] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1164.868058] env[65788]: value = "task-4663234" [ 1164.868058] env[65788]: _type = "Task" [ 1164.868058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.877874] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.905612] env[65788]: DEBUG nova.network.neutron [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Successfully created port: 4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1164.931711] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663233, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.072475] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "efe1048b-50e9-4add-910a-607a95759c7a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1165.072858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "efe1048b-50e9-4add-910a-607a95759c7a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1165.073141] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "efe1048b-50e9-4add-910a-607a95759c7a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1165.073413] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "efe1048b-50e9-4add-910a-607a95759c7a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1165.073689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "efe1048b-50e9-4add-910a-607a95759c7a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1165.076927] env[65788]: INFO nova.compute.manager [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Terminating instance [ 1165.131143] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.131143] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.189500] env[65788]: DEBUG nova.network.neutron [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1165.212396] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.212999] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.272038] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 9438ab56-1b4c-4778-a608-de319ab0ee43] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1165.284137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.284420] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1165.284674] env[65788]: DEBUG nova.network.neutron [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1165.330712] env[65788]: WARNING neutronclient.v2_0.client [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1165.331448] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.331794] env[65788]: WARNING openstack [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.355518] env[65788]: DEBUG nova.scheduler.client.report [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1165.383618] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663234, 'name': PowerOffVM_Task, 'duration_secs': 0.288705} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.389165] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.390600] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc710257-90b3-41bb-b5c4-b09465c1d488 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.414348] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718e1d48-c391-4f67-aba3-e5fa43abb558 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.434706] env[65788]: DEBUG oslo_vmware.api [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663233, 'name': PowerOnVM_Task, 'duration_secs': 1.170351} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.437734] env[65788]: DEBUG nova.network.neutron [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Updating instance_info_cache with network_info: [{"id": "98f4edb9-9923-4aa2-9ee0-e4db751e1691", "address": "fa:16:3e:62:e9:13", "network": {"id": "a402aad4-8e2b-4828-aa94-677c539f2831", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1253859523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fef7c297452947d7979e7972d6887130", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98f4edb9-99", "ovs_interfaceid": "98f4edb9-9923-4aa2-9ee0-e4db751e1691", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1165.438551] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1165.438649] env[65788]: INFO nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Took 9.30 seconds to spawn the instance on the hypervisor. [ 1165.439169] env[65788]: DEBUG nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1165.439994] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c918d814-5ed0-444e-b2e8-3b0e4761d1f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.453131] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.453433] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94d74dcb-2fd0-4961-a4ec-4d1d3eee0542 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.461643] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1165.461643] env[65788]: value = "task-4663235" [ 1165.461643] env[65788]: _type = "Task" [ 1165.461643] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.473014] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1165.473258] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1165.476849] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.476849] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1165.476849] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1165.476849] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c1b73d8-d2bb-4fa7-88ee-184ad0af88de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.484446] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1165.484645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1165.485434] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41965d08-bda3-4a0e-90c0-b447772d97e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.491495] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1165.491495] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223f27e-4bea-c075-01f2-172fb9bb7614" [ 1165.491495] env[65788]: _type = "Task" [ 1165.491495] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.501454] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223f27e-4bea-c075-01f2-172fb9bb7614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.578645] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1165.581291] env[65788]: DEBUG nova.compute.manager [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1165.581484] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1165.582389] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99630a2f-c713-453a-8160-e462e2bc9d12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.591077] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.591389] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ba07756-258d-4fdf-8bc0-9c9659180a98 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.600685] env[65788]: DEBUG oslo_vmware.api [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1165.600685] env[65788]: value = "task-4663236" [ 1165.600685] env[65788]: _type = "Task" [ 1165.600685] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.603097] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1165.603358] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1165.603518] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1165.603699] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1165.603848] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1165.603991] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1165.604252] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1165.604415] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1165.604581] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1165.604741] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1165.604913] env[65788]: DEBUG nova.virt.hardware [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1165.605858] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6cba4d-bd26-494b-94cb-8265f9e6b8a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.621665] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4e09c3-f70a-4dab-b31e-aa8f179770f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.626153] env[65788]: DEBUG oslo_vmware.api [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.777420] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f4a490a2-f2b1-4eac-8c1a-a18758583c70] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1165.788189] env[65788]: WARNING neutronclient.v2_0.client [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1165.789331] env[65788]: WARNING openstack [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.789761] env[65788]: WARNING openstack [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.931137] env[65788]: DEBUG nova.compute.manager [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1165.931389] env[65788]: DEBUG nova.compute.manager [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing instance network info cache due to event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1165.931612] env[65788]: DEBUG oslo_concurrency.lockutils [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.931754] env[65788]: DEBUG oslo_concurrency.lockutils [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1165.931938] env[65788]: DEBUG nova.network.neutron [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1165.945564] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Releasing lock "refresh_cache-22cd2208-ebb5-401a-a06d-3dd6cd591869" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.945564] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Instance network_info: |[{"id": "98f4edb9-9923-4aa2-9ee0-e4db751e1691", "address": "fa:16:3e:62:e9:13", "network": {"id": "a402aad4-8e2b-4828-aa94-677c539f2831", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1253859523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fef7c297452947d7979e7972d6887130", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98f4edb9-99", "ovs_interfaceid": "98f4edb9-9923-4aa2-9ee0-e4db751e1691", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1165.946346] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:e9:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '98f4edb9-9923-4aa2-9ee0-e4db751e1691', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1165.953884] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Creating folder: Project (fef7c297452947d7979e7972d6887130). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1165.958973] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9bb79340-a72b-4404-b68c-24ae07a44706 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.963887] env[65788]: INFO nova.compute.manager [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Took 16.72 seconds to build instance. [ 1165.972938] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Created folder: Project (fef7c297452947d7979e7972d6887130) in parent group-v910111. [ 1165.973157] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Creating folder: Instances. Parent ref: group-v910430. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1165.973414] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9439f1ca-4e3e-49df-9587-5737693c7e94 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.984664] env[65788]: WARNING openstack [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1165.985062] env[65788]: WARNING openstack [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1165.993920] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Created folder: Instances in parent group-v910430. [ 1165.994223] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1165.997962] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1165.998224] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea919fb7-38f4-4f8d-9e82-e115de773a68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.019711] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5223f27e-4bea-c075-01f2-172fb9bb7614, 'name': SearchDatastore_Task, 'duration_secs': 0.011613} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.021822] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1166.021822] env[65788]: value = "task-4663239" [ 1166.021822] env[65788]: _type = "Task" [ 1166.021822] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.022061] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4d0eee2-efc4-4e04-b2d3-b74a3a107c89 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.025724] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1166.025956] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910429', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'name': 'volume-814933a5-cdb9-4d95-b486-fd9349ad7a21', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'serial': '814933a5-cdb9-4d95-b486-fd9349ad7a21'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1166.030943] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872d9ba6-ee6d-4f90-8af1-3a96f93a03da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.038371] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1166.038371] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52aeb1eb-b500-8398-2e0c-e56172caade4" [ 1166.038371] env[65788]: _type = "Task" [ 1166.038371] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.053607] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663239, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.058793] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd9f26c-7086-47e6-96f1-e51216f12ade {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.077050] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52aeb1eb-b500-8398-2e0c-e56172caade4, 'name': SearchDatastore_Task, 'duration_secs': 0.023229} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.094201] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.094524] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. {{(pid=65788) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1166.102815] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-814933a5-cdb9-4d95-b486-fd9349ad7a21/volume-814933a5-cdb9-4d95-b486-fd9349ad7a21.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1166.103180] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ceed426-c981-4aeb-8427-9bc811e4dde8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.107108] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6132b0e-628f-4837-8fc4-a7ea0ad1415d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.125603] env[65788]: DEBUG nova.compute.manager [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Received event network-changed-98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1166.125968] env[65788]: DEBUG nova.compute.manager [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Refreshing instance network info cache due to event network-changed-98f4edb9-9923-4aa2-9ee0-e4db751e1691. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1166.126191] env[65788]: DEBUG oslo_concurrency.lockutils [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Acquiring lock "refresh_cache-22cd2208-ebb5-401a-a06d-3dd6cd591869" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.126355] env[65788]: DEBUG oslo_concurrency.lockutils [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Acquired lock "refresh_cache-22cd2208-ebb5-401a-a06d-3dd6cd591869" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.126541] env[65788]: DEBUG nova.network.neutron [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Refreshing network info cache for port 98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1166.131712] env[65788]: WARNING neutronclient.v2_0.client [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.131712] env[65788]: WARNING openstack [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.131712] env[65788]: WARNING openstack [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.143161] env[65788]: WARNING neutronclient.v2_0.client [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.143778] env[65788]: WARNING openstack [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.144137] env[65788]: WARNING openstack [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.156087] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1166.156087] env[65788]: value = "task-4663241" [ 1166.156087] env[65788]: _type = "Task" [ 1166.156087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.156829] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1166.156829] env[65788]: value = "task-4663240" [ 1166.156829] env[65788]: _type = "Task" [ 1166.156829] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.168852] env[65788]: DEBUG oslo_vmware.api [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663236, 'name': PowerOffVM_Task, 'duration_secs': 0.233141} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.170157] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1166.170394] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.170894] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bed35d9-a6df-4f92-9732-037f7f0b4127 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.180201] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663240, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.180569] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663241, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.256492] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.257045] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.257045] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleting the datastore file [datastore1] efe1048b-50e9-4add-910a-607a95759c7a {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.257396] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4c53ae6-4f73-4562-a11d-eac6dd13dbe8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.266709] env[65788]: DEBUG oslo_vmware.api [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for the task: (returnval){ [ 1166.266709] env[65788]: value = "task-4663243" [ 1166.266709] env[65788]: _type = "Task" [ 1166.266709] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.277488] env[65788]: DEBUG oslo_vmware.api [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.284743] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 1ef08e8f-744d-4939-9433-a4301cd31e77] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1166.317866] env[65788]: DEBUG nova.network.neutron [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1166.371132] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.329s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.375385] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.617s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.376196] env[65788]: DEBUG nova.objects.instance [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'resources' on Instance uuid b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.437850] env[65788]: WARNING neutronclient.v2_0.client [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.438663] env[65788]: WARNING openstack [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.439052] env[65788]: WARNING openstack [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.466862] env[65788]: DEBUG oslo_concurrency.lockutils [None req-10c7e793-21ae-421a-a517-4fae74dcefe7 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "f8c57215-ade6-424b-be8c-075998a764af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.252s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.518787] env[65788]: DEBUG nova.network.neutron [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Successfully updated port: 4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1166.537800] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663239, 'name': CreateVM_Task, 'duration_secs': 0.384212} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.538220] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1166.538930] env[65788]: WARNING neutronclient.v2_0.client [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.539623] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.539623] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.540035] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1166.541202] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8be447b7-7d31-4fdc-aeab-15a2d05fd8c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.548805] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1166.548805] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc7810-e086-2f80-e4cb-58045c143f00" [ 1166.548805] env[65788]: _type = "Task" [ 1166.548805] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.559867] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc7810-e086-2f80-e4cb-58045c143f00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.571131] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.586755] env[65788]: WARNING openstack [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.587183] env[65788]: WARNING openstack [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.681961] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663241, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.685625] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663240, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.728108] env[65788]: WARNING neutronclient.v2_0.client [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.728850] env[65788]: WARNING openstack [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.729214] env[65788]: WARNING openstack [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.777543] env[65788]: DEBUG oslo_vmware.api [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Task: {'id': task-4663243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.419343} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.777901] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.777948] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1166.778135] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1166.778433] env[65788]: INFO nova.compute.manager [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1166.778690] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1166.778883] env[65788]: DEBUG nova.compute.manager [-] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1166.778980] env[65788]: DEBUG nova.network.neutron [-] [instance: efe1048b-50e9-4add-910a-607a95759c7a] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1166.779238] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.779754] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.780480] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.789484] env[65788]: WARNING openstack [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.789925] env[65788]: WARNING openstack [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1166.797568] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 86ac6437-511a-4d72-aff0-0325e2d633f6] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1166.825519] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.873295] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.881714] env[65788]: DEBUG nova.objects.instance [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'numa_topology' on Instance uuid b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1166.908477] env[65788]: DEBUG nova.network.neutron [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Updated VIF entry in instance network info cache for port 98f4edb9-9923-4aa2-9ee0-e4db751e1691. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1166.908826] env[65788]: DEBUG nova.network.neutron [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Updating instance_info_cache with network_info: [{"id": "98f4edb9-9923-4aa2-9ee0-e4db751e1691", "address": "fa:16:3e:62:e9:13", "network": {"id": "a402aad4-8e2b-4828-aa94-677c539f2831", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1253859523-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fef7c297452947d7979e7972d6887130", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap98f4edb9-99", "ovs_interfaceid": "98f4edb9-9923-4aa2-9ee0-e4db751e1691", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1166.940524] env[65788]: INFO nova.scheduler.client.report [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocation for migration b175a0c6-bdf7-452a-b15d-6abb54724613 [ 1166.995578] env[65788]: WARNING neutronclient.v2_0.client [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1166.996252] env[65788]: WARNING openstack [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1166.996776] env[65788]: WARNING openstack [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.021980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-31853be1-f03d-4cd9-91d7-e9542dc1cadc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.021980] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-31853be1-f03d-4cd9-91d7-e9542dc1cadc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1167.022163] env[65788]: DEBUG nova.network.neutron [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1167.064531] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52dc7810-e086-2f80-e4cb-58045c143f00, 'name': SearchDatastore_Task, 'duration_secs': 0.068455} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.065455] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.066113] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1167.066113] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.066220] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1167.066386] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.067135] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41b19f3e-83c7-44b6-a660-861641506ddc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.080727] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1167.080989] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1167.081723] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fb9d457-f2c6-4e48-b2d0-e190c4354a61 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.088639] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1167.088639] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5267512d-5a21-ad29-cef9-46e649eade45" [ 1167.088639] env[65788]: _type = "Task" [ 1167.088639] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.102690] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5267512d-5a21-ad29-cef9-46e649eade45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.175630] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663241, 'name': ReconfigVM_Task, 'duration_secs': 0.973196} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.180826] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-814933a5-cdb9-4d95-b486-fd9349ad7a21/volume-814933a5-cdb9-4d95-b486-fd9349ad7a21.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1167.186712] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663240, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539434} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.189270] env[65788]: DEBUG nova.network.neutron [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updated VIF entry in instance network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1167.189650] env[65788]: DEBUG nova.network.neutron [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": null, "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1167.191587] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a05e36c-9d41-4543-89f7-790cd2e583e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.204853] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk. [ 1167.206579] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdf8c03-466e-4a5a-b246-4834267de431 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.243718] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.246556] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61fa588c-7bda-4cb4-9d6d-b48664938f1d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.260445] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1167.260445] env[65788]: value = "task-4663244" [ 1167.260445] env[65788]: _type = "Task" [ 1167.260445] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.267787] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1167.267787] env[65788]: value = "task-4663245" [ 1167.267787] env[65788]: _type = "Task" [ 1167.267787] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.275345] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663244, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.281497] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663245, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.301144] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 6215ea56-f99c-42f3-a2d5-f4d31952c8e6] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1167.330124] env[65788]: DEBUG nova.compute.manager [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1167.331071] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cf1c51-4576-438e-a406-73df93390803 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.386080] env[65788]: DEBUG nova.objects.base [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1167.413769] env[65788]: DEBUG oslo_concurrency.lockutils [req-3533d0be-a8f0-431c-974d-6b2afd8cef32 req-8274e4b8-3a91-4360-adb1-4fb420c730c1 service nova] Releasing lock "refresh_cache-22cd2208-ebb5-401a-a06d-3dd6cd591869" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.447692] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.914s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.528621] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1167.529014] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.578953] env[65788]: DEBUG nova.network.neutron [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1167.591180] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae4d065-bd46-4b2e-afd5-5fe8b71e5816 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.598625] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1167.599019] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.613185] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de822a3c-c24b-469d-9134-f0daa5add219 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.617119] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5267512d-5a21-ad29-cef9-46e649eade45, 'name': SearchDatastore_Task, 'duration_secs': 0.043344} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.618831] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d279c5c-8a18-42e1-8b50-ff7f7c894c2e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.646243] env[65788]: DEBUG nova.network.neutron [-] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1167.653929] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6120b15-9427-4be2-9e9b-b252c6decc1a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.658796] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1167.658796] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525308de-b4e2-a301-bd12-28f571aafc4a" [ 1167.658796] env[65788]: _type = "Task" [ 1167.658796] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.667034] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b15706-b3be-41d2-8754-13117b884136 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.675608] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525308de-b4e2-a301-bd12-28f571aafc4a, 'name': SearchDatastore_Task, 'duration_secs': 0.013157} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.675818] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.676259] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 22cd2208-ebb5-401a-a06d-3dd6cd591869/22cd2208-ebb5-401a-a06d-3dd6cd591869.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1167.676401] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c7c88e4-6a9d-41d9-98af-31fe9f0b8066 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.687209] env[65788]: DEBUG nova.compute.provider_tree [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.692083] env[65788]: DEBUG oslo_concurrency.lockutils [req-3efab68d-6950-4143-87ad-3fba99f9d8d9 req-7a994c02-d463-4582-b6a2-0a88a5d712c0 service nova] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.695042] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1167.695042] env[65788]: value = "task-4663246" [ 1167.695042] env[65788]: _type = "Task" [ 1167.695042] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.707324] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663246, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.709192] env[65788]: WARNING neutronclient.v2_0.client [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1167.709861] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1167.710262] env[65788]: WARNING openstack [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1167.780324] env[65788]: DEBUG oslo_vmware.api [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663244, 'name': ReconfigVM_Task, 'duration_secs': 0.291424} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.780575] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910429', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'name': 'volume-814933a5-cdb9-4d95-b486-fd9349ad7a21', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'serial': '814933a5-cdb9-4d95-b486-fd9349ad7a21'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1167.785963] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663245, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.805023] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d3e2b5f6-20f1-40ac-b92e-e110ee119178] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1167.814996] env[65788]: DEBUG nova.network.neutron [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Updating instance_info_cache with network_info: [{"id": "4c3f94f3-891e-4229-9363-1f8123b187f7", "address": "fa:16:3e:a4:52:fa", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c3f94f3-89", "ovs_interfaceid": "4c3f94f3-891e-4229-9363-1f8123b187f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1168.012521] env[65788]: DEBUG nova.compute.manager [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Received event network-vif-plugged-4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1168.012797] env[65788]: DEBUG oslo_concurrency.lockutils [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Acquiring lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.013186] env[65788]: DEBUG oslo_concurrency.lockutils [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.013391] env[65788]: DEBUG oslo_concurrency.lockutils [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.013546] env[65788]: DEBUG nova.compute.manager [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] No waiting events found dispatching network-vif-plugged-4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1168.013713] env[65788]: WARNING nova.compute.manager [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Received unexpected event network-vif-plugged-4c3f94f3-891e-4229-9363-1f8123b187f7 for instance with vm_state building and task_state spawning. [ 1168.013925] env[65788]: DEBUG nova.compute.manager [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Received event network-changed-4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1168.014126] env[65788]: DEBUG nova.compute.manager [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Refreshing instance network info cache due to event network-changed-4c3f94f3-891e-4229-9363-1f8123b187f7. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1168.014375] env[65788]: DEBUG oslo_concurrency.lockutils [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Acquiring lock "refresh_cache-31853be1-f03d-4cd9-91d7-e9542dc1cadc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.154031] env[65788]: INFO nova.compute.manager [-] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Took 1.37 seconds to deallocate network for instance. [ 1168.191202] env[65788]: DEBUG nova.scheduler.client.report [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1168.199327] env[65788]: DEBUG nova.compute.manager [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-changed-fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1168.199327] env[65788]: DEBUG nova.compute.manager [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing instance network info cache due to event network-changed-fda83fac-56b6-4ab0-824a-7af792b7c8c7. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1168.199327] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.199327] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.199327] env[65788]: DEBUG nova.network.neutron [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing network info cache for port fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1168.211857] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663246, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.286528] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663245, 'name': ReconfigVM_Task, 'duration_secs': 0.608009} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.287320] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Reconfigured VM instance instance-00000070 to attach disk [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b-rescue.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.288169] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ceee6a-1016-4077-a1f2-de3c1a789617 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.312043] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5cef20ed-d9f0-4237-ae9d-da401b953904] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1168.320068] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-31853be1-f03d-4cd9-91d7-e9542dc1cadc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1168.320311] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Instance network_info: |[{"id": "4c3f94f3-891e-4229-9363-1f8123b187f7", "address": "fa:16:3e:a4:52:fa", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c3f94f3-89", "ovs_interfaceid": "4c3f94f3-891e-4229-9363-1f8123b187f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1168.320663] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-608d4cf5-b630-4fff-905c-5c9ab99ea62c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.333991] env[65788]: DEBUG oslo_concurrency.lockutils [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Acquired lock "refresh_cache-31853be1-f03d-4cd9-91d7-e9542dc1cadc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.334717] env[65788]: DEBUG nova.network.neutron [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Refreshing network info cache for port 4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1168.335802] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:52:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c3f94f3-891e-4229-9363-1f8123b187f7', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1168.343331] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1168.344385] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1168.345693] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d562c6b-3863-4657-a7c3-b54a4e0f7ec1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.365722] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1168.365722] env[65788]: value = "task-4663247" [ 1168.365722] env[65788]: _type = "Task" [ 1168.365722] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.366521] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7602b06d-39a8-4738-9011-0845f1f324ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.378120] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1168.378120] env[65788]: value = "task-4663248" [ 1168.378120] env[65788]: _type = "Task" [ 1168.378120] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.388027] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663247, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.388171] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Doing hard reboot of VM {{(pid=65788) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1168.388794] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-4691c546-16fb-4a4e-90e6-9df265238c52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.394795] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663248, 'name': CreateVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.396821] env[65788]: DEBUG oslo_vmware.api [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1168.396821] env[65788]: value = "task-4663249" [ 1168.396821] env[65788]: _type = "Task" [ 1168.396821] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.408188] env[65788]: DEBUG oslo_vmware.api [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663249, 'name': ResetVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.555713] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.556084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.556381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.556676] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.556927] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.559292] env[65788]: INFO nova.compute.manager [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Terminating instance [ 1168.663075] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.709094] env[65788]: WARNING neutronclient.v2_0.client [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.710094] env[65788]: WARNING openstack [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.710534] env[65788]: WARNING openstack [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.718576] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.343s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1168.721264] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663246, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.580179} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.721830] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.059s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1168.722084] env[65788]: DEBUG nova.objects.instance [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lazy-loading 'resources' on Instance uuid efe1048b-50e9-4add-910a-607a95759c7a {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.723344] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 22cd2208-ebb5-401a-a06d-3dd6cd591869/22cd2208-ebb5-401a-a06d-3dd6cd591869.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1168.723557] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1168.723965] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d31ceceb-18a6-40c0-9e35-d6ac3cc11bb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.733053] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1168.733053] env[65788]: value = "task-4663250" [ 1168.733053] env[65788]: _type = "Task" [ 1168.733053] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.745035] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.820834] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: eceafff8-7d28-4b9b-ade6-5736d5977efa] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1168.845592] env[65788]: DEBUG nova.objects.instance [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.848031] env[65788]: WARNING neutronclient.v2_0.client [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.848445] env[65788]: WARNING openstack [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.848815] env[65788]: WARNING openstack [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.860254] env[65788]: WARNING openstack [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.860642] env[65788]: WARNING openstack [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.882019] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663247, 'name': ReconfigVM_Task, 'duration_secs': 0.306732} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.885360] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1168.888075] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-86d5df05-2dec-466e-ba28-e95f2324ad71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.897588] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663248, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.902803] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1168.902803] env[65788]: value = "task-4663251" [ 1168.902803] env[65788]: _type = "Task" [ 1168.902803] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.915218] env[65788]: DEBUG oslo_vmware.api [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663249, 'name': ResetVM_Task, 'duration_secs': 0.132811} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.916198] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Did hard reboot of VM {{(pid=65788) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1168.916454] env[65788]: DEBUG nova.compute.manager [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1168.917347] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68598c29-de04-4cc5-8641-e0002401eb90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.923833] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663251, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.956514] env[65788]: WARNING openstack [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.956850] env[65788]: WARNING openstack [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1168.968609] env[65788]: WARNING neutronclient.v2_0.client [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1168.969278] env[65788]: WARNING openstack [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1168.969716] env[65788]: WARNING openstack [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.023128] env[65788]: WARNING neutronclient.v2_0.client [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.023817] env[65788]: WARNING openstack [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1169.024246] env[65788]: WARNING openstack [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1169.059584] env[65788]: DEBUG nova.network.neutron [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updated VIF entry in instance network info cache for port fda83fac-56b6-4ab0-824a-7af792b7c8c7. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1169.060103] env[65788]: DEBUG nova.network.neutron [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1169.063863] env[65788]: DEBUG nova.compute.manager [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1169.064341] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1169.065565] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44554e6a-2037-4873-ac8d-205c0302b817 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.077598] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.077996] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95dddb78-d089-4d2f-a30a-a62e2d92ddcc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.090508] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1169.090508] env[65788]: value = "task-4663252" [ 1169.090508] env[65788]: _type = "Task" [ 1169.090508] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.112704] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.140297] env[65788]: DEBUG nova.network.neutron [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Updated VIF entry in instance network info cache for port 4c3f94f3-891e-4229-9363-1f8123b187f7. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1169.140841] env[65788]: DEBUG nova.network.neutron [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Updating instance_info_cache with network_info: [{"id": "4c3f94f3-891e-4229-9363-1f8123b187f7", "address": "fa:16:3e:a4:52:fa", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c3f94f3-89", "ovs_interfaceid": "4c3f94f3-891e-4229-9363-1f8123b187f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1169.230962] env[65788]: DEBUG oslo_concurrency.lockutils [None req-17d45daf-d640-414f-bf23-8a158379a700 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.941s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.232218] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.661s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.232929] env[65788]: INFO nova.compute.manager [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Unshelving [ 1169.244720] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076735} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.245751] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1169.246808] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7adf4f-0b45-4803-8d4b-8bd72c5d2140 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.276141] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 22cd2208-ebb5-401a-a06d-3dd6cd591869/22cd2208-ebb5-401a-a06d-3dd6cd591869.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1169.279465] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61712834-be81-43f9-8422-0ec3ea92669f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.301407] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1169.301407] env[65788]: value = "task-4663253" [ 1169.301407] env[65788]: _type = "Task" [ 1169.301407] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.310384] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.327798] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: ed55713e-8bf1-4960-8309-47498b163d02] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1169.362747] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbf0829d-ad3d-4ca9-a280-da52dd98f5b4 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.954s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.396971] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663248, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.416379] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663251, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.436097] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d57d38e-4e29-45e5-9f9b-ea0482afaddb tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.676s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.479337] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4737beeb-386a-4f19-aaa7-c4d35f7295e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.490664] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756b02a9-b25a-4f75-9d81-deb19ba81f5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.531389] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502ecc45-078f-464b-adeb-39252cd0cfef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.543151] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1beb40-0e78-4661-8fb0-eef67710ae1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.560645] env[65788]: DEBUG nova.compute.provider_tree [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.563603] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1169.563855] env[65788]: DEBUG nova.compute.manager [req-f1b65d9d-18f2-4a9c-82ee-0bbf060d8a3f req-4ae9e7d0-2fa1-4838-9e30-c237fcdac3a5 service nova] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Received event network-vif-deleted-e936d0e7-de51-4ac2-9b4a-08f999624c5c {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1169.609407] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663252, 'name': PowerOffVM_Task, 'duration_secs': 0.3845} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.609786] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1169.610083] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1169.610364] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8f37d5d-0d1f-4ab8-adac-bfafd35a4010 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.647593] env[65788]: DEBUG oslo_concurrency.lockutils [req-20ecf0e7-83eb-4042-9228-bd182a438db5 req-6f8cd4e1-9260-4c3a-a5ac-6591a66b51fa service nova] Releasing lock "refresh_cache-31853be1-f03d-4cd9-91d7-e9542dc1cadc" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1169.778460] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1169.778790] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1169.779051] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleting the datastore file [datastore1] 5621bbe8-131f-4266-9109-a3be445310a0 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1169.779767] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b54e60e-ca8a-4887-b895-779fea4b6823 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.788376] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for the task: (returnval){ [ 1169.788376] env[65788]: value = "task-4663255" [ 1169.788376] env[65788]: _type = "Task" [ 1169.788376] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.798280] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.810938] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663253, 'name': ReconfigVM_Task, 'duration_secs': 0.342481} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.811230] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 22cd2208-ebb5-401a-a06d-3dd6cd591869/22cd2208-ebb5-401a-a06d-3dd6cd591869.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1169.811915] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50bf6c1d-b6b7-459f-81fd-3d5bdf6fca2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.819570] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1169.819570] env[65788]: value = "task-4663256" [ 1169.819570] env[65788]: _type = "Task" [ 1169.819570] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.829260] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663256, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.831893] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.832105] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Cleaning up deleted instances with incomplete migration {{(pid=65788) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11954}} [ 1169.894190] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663248, 'name': CreateVM_Task, 'duration_secs': 1.511378} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.894445] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1169.895048] env[65788]: WARNING neutronclient.v2_0.client [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1169.895497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.895700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.896085] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1169.896388] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f53a5289-e25e-47d8-97ed-f1c7d04188ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.903058] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1169.903058] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52440029-a3dd-aadf-f14d-a38cd3cbb584" [ 1169.903058] env[65788]: _type = "Task" [ 1169.903058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.915730] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52440029-a3dd-aadf-f14d-a38cd3cbb584, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.919329] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663251, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.062766] env[65788]: DEBUG nova.scheduler.client.report [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1170.147418] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.147742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1170.271164] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.300322] env[65788]: DEBUG oslo_vmware.api [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Task: {'id': task-4663255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294475} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.300794] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.300890] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1170.301132] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1170.301387] env[65788]: INFO nova.compute.manager [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1170.301733] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1170.301959] env[65788]: DEBUG nova.compute.manager [-] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1170.302095] env[65788]: DEBUG nova.network.neutron [-] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1170.302423] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.303104] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1170.303458] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1170.332831] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663256, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.379156] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1170.419282] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52440029-a3dd-aadf-f14d-a38cd3cbb584, 'name': SearchDatastore_Task, 'duration_secs': 0.015055} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.422020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.422326] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1170.422599] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.422866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1170.422959] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1170.423311] env[65788]: DEBUG oslo_vmware.api [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663251, 'name': PowerOnVM_Task, 'duration_secs': 1.413914} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.423541] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-892a7442-9370-43d0-8358-f6c05367321e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.425801] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1170.429181] env[65788]: DEBUG nova.compute.manager [None req-cde2fe92-4b67-4757-ad0f-512afca2fc1c tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1170.430335] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1713c1f8-db43-44b1-8d83-4f52b459974d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.435943] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1170.436195] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1170.438412] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53414299-aa24-4251-8e31-bbf25de5a301 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.447086] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1170.447086] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f499e2-9559-14f5-874d-8537e3a90cdf" [ 1170.447086] env[65788]: _type = "Task" [ 1170.447086] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.456782] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f499e2-9559-14f5-874d-8537e3a90cdf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.568869] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.573820] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.303s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1170.574287] env[65788]: DEBUG nova.objects.instance [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'pci_requests' on Instance uuid b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.594413] env[65788]: INFO nova.scheduler.client.report [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Deleted allocations for instance efe1048b-50e9-4add-910a-607a95759c7a [ 1170.600144] env[65788]: DEBUG nova.compute.manager [req-5e590a46-b824-42c8-abf2-bdad0af83ca5 req-f296ab74-1fec-413e-b9df-250774ba362b service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Received event network-vif-deleted-ec1870b5-8b77-4ad8-9779-fdcd725dd800 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1170.600351] env[65788]: INFO nova.compute.manager [req-5e590a46-b824-42c8-abf2-bdad0af83ca5 req-f296ab74-1fec-413e-b9df-250774ba362b service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Neutron deleted interface ec1870b5-8b77-4ad8-9779-fdcd725dd800; detaching it from the instance and deleting it from the info cache [ 1170.600518] env[65788]: DEBUG nova.network.neutron [req-5e590a46-b824-42c8-abf2-bdad0af83ca5 req-f296ab74-1fec-413e-b9df-250774ba362b service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1170.650757] env[65788]: INFO nova.compute.manager [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Detaching volume ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b [ 1170.689052] env[65788]: INFO nova.virt.block_device [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Attempting to driver detach volume ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b from mountpoint /dev/sdb [ 1170.689322] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1170.689505] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910426', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'name': 'volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'serial': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1170.690673] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733d299f-bfc0-4377-bee2-fe215ede8a4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.722199] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b00a75-bf60-44c3-be2a-1a338c214be5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.733601] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69161b2f-36f0-44ff-94a7-1dfef48a3cbf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.760046] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f5a1d9-8eb7-46f3-a5e7-5f7f9a528170 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.776876] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The volume has not been displaced from its original location: [datastore2] volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b/volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1170.782208] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1170.782597] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-636c2fb8-9f15-4009-8e91-811e63f13527 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.802373] env[65788]: DEBUG oslo_vmware.api [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1170.802373] env[65788]: value = "task-4663257" [ 1170.802373] env[65788]: _type = "Task" [ 1170.802373] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.811268] env[65788]: DEBUG oslo_vmware.api [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663257, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.831336] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663256, 'name': Rename_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.959506] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f499e2-9559-14f5-874d-8537e3a90cdf, 'name': SearchDatastore_Task, 'duration_secs': 0.03257} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.960405] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3258ab8-121c-4db5-a38d-266eed7a7a2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.967981] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1170.967981] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287610d-b0f9-334f-3f56-69240d891e96" [ 1170.967981] env[65788]: _type = "Task" [ 1170.967981] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.980092] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287610d-b0f9-334f-3f56-69240d891e96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.079031] env[65788]: DEBUG nova.objects.instance [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'numa_topology' on Instance uuid b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.082774] env[65788]: DEBUG nova.network.neutron [-] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1171.106272] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0eadc55d-1565-44f1-92e9-71ed5d12f247 tempest-ServerActionsTestOtherA-298845163 tempest-ServerActionsTestOtherA-298845163-project-member] Lock "efe1048b-50e9-4add-910a-607a95759c7a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.033s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.107354] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92e218c7-f570-4cf9-bddd-649f2bc8fd8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.119362] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b1592a-28fc-4b2f-ac88-ea116fb87a96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.168252] env[65788]: DEBUG nova.compute.manager [req-5e590a46-b824-42c8-abf2-bdad0af83ca5 req-f296ab74-1fec-413e-b9df-250774ba362b service nova] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Detach interface failed, port_id=ec1870b5-8b77-4ad8-9779-fdcd725dd800, reason: Instance 5621bbe8-131f-4266-9109-a3be445310a0 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1171.313694] env[65788]: DEBUG oslo_vmware.api [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663257, 'name': ReconfigVM_Task, 'duration_secs': 0.431917} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.313949] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1171.318763] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1de495f4-09e4-4aa1-8289-e443867ea2f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.329403] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1171.340434] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663256, 'name': Rename_Task, 'duration_secs': 1.159519} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.342246] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1171.342536] env[65788]: DEBUG oslo_vmware.api [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1171.342536] env[65788]: value = "task-4663258" [ 1171.342536] env[65788]: _type = "Task" [ 1171.342536] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.342761] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6717e56f-1362-4f68-950d-a93d3b3fa2f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.352475] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1171.352475] env[65788]: value = "task-4663259" [ 1171.352475] env[65788]: _type = "Task" [ 1171.352475] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.357084] env[65788]: DEBUG oslo_vmware.api [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663258, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.368904] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663259, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.481424] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5287610d-b0f9-334f-3f56-69240d891e96, 'name': SearchDatastore_Task, 'duration_secs': 0.012074} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.481424] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1171.481533] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 31853be1-f03d-4cd9-91d7-e9542dc1cadc/31853be1-f03d-4cd9-91d7-e9542dc1cadc.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1171.481863] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a1a5824-95a3-4941-b826-0c5d2d270db2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.491828] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1171.491828] env[65788]: value = "task-4663260" [ 1171.491828] env[65788]: _type = "Task" [ 1171.491828] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.502463] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.583630] env[65788]: INFO nova.compute.claims [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1171.587142] env[65788]: INFO nova.compute.manager [-] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Took 1.28 seconds to deallocate network for instance. [ 1171.841940] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_power_states {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1171.869034] env[65788]: DEBUG oslo_vmware.api [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663258, 'name': ReconfigVM_Task, 'duration_secs': 0.236119} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.869034] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910426', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'name': 'volume-ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b', 'serial': 'ab24d355-2e7b-41fb-ab0c-0b78e6c53a8b'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1171.876228] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663259, 'name': PowerOnVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.913416] env[65788]: INFO nova.compute.manager [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Unrescuing [ 1171.913416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.913416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquired lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1171.913416] env[65788]: DEBUG nova.network.neutron [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1172.003576] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663260, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.095979] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.359137] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Getting list of instances from cluster (obj){ [ 1172.359137] env[65788]: value = "domain-c8" [ 1172.359137] env[65788]: _type = "ClusterComputeResource" [ 1172.359137] env[65788]: } {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1172.361286] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8d23a5-ec59-45de-a411-26f949722244 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.378602] env[65788]: DEBUG oslo_vmware.api [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663259, 'name': PowerOnVM_Task, 'duration_secs': 0.570402} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.389532] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1172.389866] env[65788]: INFO nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Took 9.14 seconds to spawn the instance on the hypervisor. [ 1172.390124] env[65788]: DEBUG nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1172.390502] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Got total of 10 instances {{(pid=65788) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1172.390695] env[65788]: WARNING nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] While synchronizing instance power states, found 12 instances in the database and 10 instances on the hypervisor. [ 1172.390873] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.391239] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.391532] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.391797] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 27ed2344-c42b-46bc-b51a-20821f67cbf0 {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.392123] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 6012182d-8b4e-4379-949f-d3a26211699b {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.392341] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 5621bbe8-131f-4266-9109-a3be445310a0 {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.392610] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 50c70e36-a3dc-4e8c-8fe6-74039d1828bf {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.392893] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.393165] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid f59d609a-e5bb-4202-ac05-d486bbf9e05b {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.393434] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.393693] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 22cd2208-ebb5-401a-a06d-3dd6cd591869 {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.393951] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Triggering sync for uuid 31853be1-f03d-4cd9-91d7-e9542dc1cadc {{(pid=65788) _sync_power_states /opt/stack/nova/nova/compute/manager.py:11027}} [ 1172.394928] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c20fb56-c5e1-45f4-bf84-ef61877d5dad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.397758] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.398049] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.398392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.398679] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.398940] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.399210] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.399524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "6012182d-8b4e-4379-949f-d3a26211699b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.399763] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "6012182d-8b4e-4379-949f-d3a26211699b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.400092] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "5621bbe8-131f-4266-9109-a3be445310a0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.400377] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.400614] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.401192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.401381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "bdd09116-31de-491d-b129-c117f898881e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.401626] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.401820] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.401980] env[65788]: INFO nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] During sync_power_state the instance has a pending task (unrescuing). Skip. [ 1172.402149] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.402427] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.402607] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "f8c57215-ade6-424b-be8c-075998a764af" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.402849] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.403136] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.404064] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94353f0-d4fa-4b26-8d8b-78f0c6d144bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.408694] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052fb1ff-171d-4a64-966f-8827052faaec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.413297] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f3baef-72ac-4f59-ad3e-17f1194095b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.417992] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3abbd838-2ed3-4dc8-aa69-e11a5365a7dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.422466] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48e4307-04c8-400b-ab3a-cf4dd855503a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.427688] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab04cdd-4ade-4ac3-aa86-c42aa8a4d128 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.434855] env[65788]: WARNING neutronclient.v2_0.client [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1172.435824] env[65788]: WARNING openstack [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.436372] env[65788]: WARNING openstack [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.450665] env[65788]: DEBUG nova.objects.instance [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.505023] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660495} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.505023] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 31853be1-f03d-4cd9-91d7-e9542dc1cadc/31853be1-f03d-4cd9-91d7-e9542dc1cadc.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1172.505271] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1172.505459] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e5af194-a924-4d23-bb17-0d6a3fee3c3a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.518435] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1172.518435] env[65788]: value = "task-4663261" [ 1172.518435] env[65788]: _type = "Task" [ 1172.518435] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.530815] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.596727] env[65788]: WARNING openstack [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.597139] env[65788]: WARNING openstack [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.687805] env[65788]: WARNING neutronclient.v2_0.client [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1172.689311] env[65788]: WARNING openstack [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1172.689311] env[65788]: WARNING openstack [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1172.809357] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4fe5bb-3dae-4276-a251-9f6d66af3a24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.821779] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39dc4594-53e3-4762-aed4-8560ffbb96b7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.832571] env[65788]: DEBUG nova.network.neutron [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Updating instance_info_cache with network_info: [{"id": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "address": "fa:16:3e:9e:1f:ca", "network": {"id": "f2c05915-1d43-4a7d-92d7-05844b0bf97d", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1255063227-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "e452e2f394f549c793ecfb4f2d5cfa57", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap458887b5-5e", "ovs_interfaceid": "458887b5-5ecd-41a3-9bb5-4ee7926f28ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1172.871034] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e8da49-f9e6-4e03-997b-dd6ad51a4d6a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.884807] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9e0f1b-9c3b-493f-81f6-c44707089f1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.902322] env[65788]: DEBUG nova.compute.provider_tree [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1172.985773] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.587s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.992031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.590s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.992031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.589s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.992031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "6012182d-8b4e-4379-949f-d3a26211699b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.590s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.992031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "bdd09116-31de-491d-b129-c117f898881e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.589s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.992031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "f8c57215-ade6-424b-be8c-075998a764af" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.588s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.992031] env[65788]: INFO nova.compute.manager [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Took 14.90 seconds to build instance. [ 1173.030031] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108212} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.030031] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1173.030653] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664374d4-f5ee-455d-a3ce-f7ab21423da2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.060594] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 31853be1-f03d-4cd9-91d7-e9542dc1cadc/31853be1-f03d-4cd9-91d7-e9542dc1cadc.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1173.061289] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5273af4e-3e66-4742-85a2-27a5280dc967 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.084558] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1173.084558] env[65788]: value = "task-4663262" [ 1173.084558] env[65788]: _type = "Task" [ 1173.084558] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.094137] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.336392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Releasing lock "refresh_cache-f59d609a-e5bb-4202-ac05-d486bbf9e05b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.337199] env[65788]: DEBUG nova.objects.instance [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lazy-loading 'flavor' on Instance uuid f59d609a-e5bb-4202-ac05-d486bbf9e05b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.405777] env[65788]: DEBUG nova.scheduler.client.report [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1173.476837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2f0867e6-2ac0-478e-bd9c-97cea1c0467b tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.329s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.478115] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.079s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.479118] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76824e67-4fe9-4b52-828e-ced222ccf836 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.493600] env[65788]: DEBUG oslo_concurrency.lockutils [None req-537fd61e-8d41-4f93-bd66-1511ae89c60c tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.413s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.494051] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.091s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.494423] env[65788]: INFO nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] During sync_power_state the instance has a pending task (spawning). Skip. [ 1173.494522] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.503825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.596090] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663262, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.843740] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb1c88d-e0ab-4cb9-ac23-d4878dbe1778 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.866663] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.867056] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-863af231-4079-4810-a69a-9973315b8eff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.876777] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1173.876777] env[65788]: value = "task-4663263" [ 1173.876777] env[65788]: _type = "Task" [ 1173.876777] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.886437] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.912943] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.338s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.912943] env[65788]: WARNING neutronclient.v2_0.client [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1173.915571] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.820s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.915828] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.957917] env[65788]: INFO nova.scheduler.client.report [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Deleted allocations for instance 5621bbe8-131f-4266-9109-a3be445310a0 [ 1173.987678] env[65788]: INFO nova.network.neutron [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating port dc5727b9-f909-4ed4-80d0-06ba5986f0cc with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1173.992811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.514s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1173.993118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.489s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.097361] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663262, 'name': ReconfigVM_Task, 'duration_secs': 0.937173} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.097921] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 31853be1-f03d-4cd9-91d7-e9542dc1cadc/31853be1-f03d-4cd9-91d7-e9542dc1cadc.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.098435] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ae02986-ac17-4066-8a34-33f5256e11c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.106812] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1174.106812] env[65788]: value = "task-4663264" [ 1174.106812] env[65788]: _type = "Task" [ 1174.106812] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.116695] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663264, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.148570] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.148951] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.149270] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "22cd2208-ebb5-401a-a06d-3dd6cd591869-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.149754] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.149986] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.153267] env[65788]: INFO nova.compute.manager [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Terminating instance [ 1174.387162] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663263, 'name': PowerOffVM_Task, 'duration_secs': 0.332777} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.387488] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.393056] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1174.393384] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66f4c8e1-320b-437c-9cda-32dcc76200e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.413288] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1174.413288] env[65788]: value = "task-4663265" [ 1174.413288] env[65788]: _type = "Task" [ 1174.413288] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.422584] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663265, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.469701] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2dd744ca-26e3-45ab-bc9d-9b492ebfa358 tempest-DeleteServersTestJSON-1456685968 tempest-DeleteServersTestJSON-1456685968-project-member] Lock "5621bbe8-131f-4266-9109-a3be445310a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.913s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.470847] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "5621bbe8-131f-4266-9109-a3be445310a0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.071s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.471394] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18636f0a-186d-4080-a094-a99dd8464344 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.482759] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc50e60-abf8-46e6-bd1c-166614a75a50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.496635] env[65788]: INFO nova.compute.manager [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Detaching volume 814933a5-cdb9-4d95-b486-fd9349ad7a21 [ 1174.541896] env[65788]: INFO nova.virt.block_device [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Attempting to driver detach volume 814933a5-cdb9-4d95-b486-fd9349ad7a21 from mountpoint /dev/sdc [ 1174.542331] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1174.542608] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910429', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'name': 'volume-814933a5-cdb9-4d95-b486-fd9349ad7a21', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'serial': '814933a5-cdb9-4d95-b486-fd9349ad7a21'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1174.543574] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a870c9e9-845a-47fc-8c82-c7b5bd81e969 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.569808] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb5ed0d4-132d-4a7f-aa46-164a9d2fcfd1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.578894] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081644bb-3583-47ef-9077-c7ff4f4c95ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.603537] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b864b80-1a0d-40c2-b3bc-b28c5f6faef2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.622449] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] The volume has not been displaced from its original location: [datastore2] volume-814933a5-cdb9-4d95-b486-fd9349ad7a21/volume-814933a5-cdb9-4d95-b486-fd9349ad7a21.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1174.628019] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfiguring VM instance instance-00000068 to detach disk 2002 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1174.628955] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3b3a7c7-7a74-4a01-9d7c-5412a86383e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.645833] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663264, 'name': Rename_Task, 'duration_secs': 0.177307} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.646760] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1174.647045] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9287349-e1ad-4bbd-81c1-2453a98d0f05 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.653086] env[65788]: DEBUG oslo_vmware.api [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1174.653086] env[65788]: value = "task-4663266" [ 1174.653086] env[65788]: _type = "Task" [ 1174.653086] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.658730] env[65788]: DEBUG nova.compute.manager [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1174.658948] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.659335] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1174.659335] env[65788]: value = "task-4663267" [ 1174.659335] env[65788]: _type = "Task" [ 1174.659335] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.660458] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51de752-3b49-48b7-8c9c-e922e04f6995 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.671157] env[65788]: DEBUG oslo_vmware.api [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663266, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.683411] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663267, 'name': PowerOnVM_Task} progress is 33%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.687140] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1174.687587] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dca931fe-2ae7-4cc1-9e9b-0c81ff8d1b3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.695542] env[65788]: DEBUG oslo_vmware.api [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1174.695542] env[65788]: value = "task-4663268" [ 1174.695542] env[65788]: _type = "Task" [ 1174.695542] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.706954] env[65788]: DEBUG oslo_vmware.api [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663268, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.927934] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663265, 'name': ReconfigVM_Task, 'duration_secs': 0.304808} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.928400] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1174.928620] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1174.928951] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fcc6fb0-366c-4177-82c2-df393a622e58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.938515] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1174.938515] env[65788]: value = "task-4663269" [ 1174.938515] env[65788]: _type = "Task" [ 1174.938515] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.949941] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.023727] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "5621bbe8-131f-4266-9109-a3be445310a0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.553s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.168584] env[65788]: DEBUG oslo_vmware.api [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663266, 'name': ReconfigVM_Task, 'duration_secs': 0.311748} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.173215] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Reconfigured VM instance instance-00000068 to detach disk 2002 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1175.179628] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a98485ea-9050-42ee-a9a7-9f31dabcf7d3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.205098] env[65788]: DEBUG oslo_vmware.api [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1175.205098] env[65788]: value = "task-4663270" [ 1175.205098] env[65788]: _type = "Task" [ 1175.205098] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.205466] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663267, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.213354] env[65788]: DEBUG oslo_vmware.api [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663268, 'name': PowerOffVM_Task, 'duration_secs': 0.25355} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.214381] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1175.214719] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1175.215095] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a494c75-a2a8-4f47-b734-5ccc42d8fa78 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.223330] env[65788]: DEBUG oslo_vmware.api [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663270, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.307496] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1175.307725] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1175.307926] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Deleting the datastore file [datastore2] 22cd2208-ebb5-401a-a06d-3dd6cd591869 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1175.308499] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-deb90781-24fd-4dc7-9967-0f744a25907e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.318226] env[65788]: DEBUG oslo_vmware.api [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for the task: (returnval){ [ 1175.318226] env[65788]: value = "task-4663272" [ 1175.318226] env[65788]: _type = "Task" [ 1175.318226] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.329318] env[65788]: DEBUG oslo_vmware.api [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663272, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.454697] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663269, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.669493] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.669684] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.669856] env[65788]: DEBUG nova.network.neutron [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1175.689143] env[65788]: DEBUG oslo_vmware.api [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663267, 'name': PowerOnVM_Task, 'duration_secs': 0.757139} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.689307] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1175.689536] env[65788]: INFO nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Took 10.11 seconds to spawn the instance on the hypervisor. [ 1175.689727] env[65788]: DEBUG nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1175.690839] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fd320d-07f6-4c75-b741-743a4d366a23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.724142] env[65788]: DEBUG oslo_vmware.api [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663270, 'name': ReconfigVM_Task, 'duration_secs': 0.254821} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.724564] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910429', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'name': 'volume-814933a5-cdb9-4d95-b486-fd9349ad7a21', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9946cef4-b8c7-41b4-9299-eace1929bf3d', 'attached_at': '', 'detached_at': '', 'volume_id': '814933a5-cdb9-4d95-b486-fd9349ad7a21', 'serial': '814933a5-cdb9-4d95-b486-fd9349ad7a21'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1175.832321] env[65788]: DEBUG oslo_vmware.api [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Task: {'id': task-4663272, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274547} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.832609] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.835027] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.835027] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.835027] env[65788]: INFO nova.compute.manager [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1175.835027] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1175.835027] env[65788]: DEBUG nova.compute.manager [-] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1175.835027] env[65788]: DEBUG nova.network.neutron [-] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1175.835027] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.835027] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1175.835525] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1175.858038] env[65788]: DEBUG nova.compute.manager [req-3781e8ab-e823-4078-bc65-3e3aa27314fc req-39ab8de0-990d-4cee-bbba-79928d2e610a service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-vif-plugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1175.858038] env[65788]: DEBUG oslo_concurrency.lockutils [req-3781e8ab-e823-4078-bc65-3e3aa27314fc req-39ab8de0-990d-4cee-bbba-79928d2e610a service nova] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.858038] env[65788]: DEBUG oslo_concurrency.lockutils [req-3781e8ab-e823-4078-bc65-3e3aa27314fc req-39ab8de0-990d-4cee-bbba-79928d2e610a service nova] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.858038] env[65788]: DEBUG oslo_concurrency.lockutils [req-3781e8ab-e823-4078-bc65-3e3aa27314fc req-39ab8de0-990d-4cee-bbba-79928d2e610a service nova] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.858038] env[65788]: DEBUG nova.compute.manager [req-3781e8ab-e823-4078-bc65-3e3aa27314fc req-39ab8de0-990d-4cee-bbba-79928d2e610a service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] No waiting events found dispatching network-vif-plugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1175.858038] env[65788]: WARNING nova.compute.manager [req-3781e8ab-e823-4078-bc65-3e3aa27314fc req-39ab8de0-990d-4cee-bbba-79928d2e610a service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received unexpected event network-vif-plugged-dc5727b9-f909-4ed4-80d0-06ba5986f0cc for instance with vm_state shelved_offloaded and task_state spawning. [ 1175.947062] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1175.954979] env[65788]: DEBUG oslo_vmware.api [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663269, 'name': PowerOnVM_Task, 'duration_secs': 0.544123} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.954979] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1175.954979] env[65788]: DEBUG nova.compute.manager [None req-eb21fa46-188a-4c56-8565-941fdedcb20b tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1175.959194] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd81314-1849-4848-9729-e80332c9e854 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.176883] env[65788]: WARNING neutronclient.v2_0.client [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.177955] env[65788]: WARNING openstack [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.178250] env[65788]: WARNING openstack [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.217489] env[65788]: INFO nova.compute.manager [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Took 17.37 seconds to build instance. [ 1176.291029] env[65788]: DEBUG nova.objects.instance [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'flavor' on Instance uuid 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.354670] env[65788]: WARNING openstack [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.355599] env[65788]: WARNING openstack [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.429445] env[65788]: WARNING neutronclient.v2_0.client [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1176.429445] env[65788]: WARNING openstack [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1176.429445] env[65788]: WARNING openstack [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1176.537123] env[65788]: DEBUG nova.network.neutron [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1176.720700] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e05db2f8-365f-4c1d-8f9c-75162a787a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.906s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.721084] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.318s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.721289] env[65788]: INFO nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] During sync_power_state the instance has a pending task (spawning). Skip. [ 1176.721457] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.733154] env[65788]: DEBUG nova.network.neutron [-] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1177.041467] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.079062] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e5c09e2934c57da87f67927e8125f12f',container_format='bare',created_at=2025-11-21T13:55:07Z,direct_url=,disk_format='vmdk',id=aa3a5ddd-40cd-4c98-8384-9c40c7390261,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1822720479-shelved',owner='2b70d3c8627449eaa6372ebe3bd90233',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2025-11-21T13:55:24Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1177.079478] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1177.079478] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1177.079746] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1177.079912] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1177.080104] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1177.080621] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1177.080621] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1177.080737] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1177.080802] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1177.081049] env[65788]: DEBUG nova.virt.hardware [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1177.082373] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70db65ff-764b-4cf9-b724-ca8bc0ceaae0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.094816] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd20a31-73ab-4387-9382-65ab6403f811 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.111814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:12:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc5727b9-f909-4ed4-80d0-06ba5986f0cc', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.120974] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1177.121128] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1177.121333] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9750e4be-846b-46e8-bb5f-1aca0ef48d55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.143682] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.143682] env[65788]: value = "task-4663274" [ 1177.143682] env[65788]: _type = "Task" [ 1177.143682] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.155480] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663274, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.236589] env[65788]: INFO nova.compute.manager [-] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Took 1.40 seconds to deallocate network for instance. [ 1177.302430] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3b3182b6-45b9-4b10-8c26-573ff40682ef tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.309s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.482029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.482029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.482029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.482029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.482029] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.483397] env[65788]: INFO nova.compute.manager [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Terminating instance [ 1177.654847] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663274, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.747641] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.747928] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.748169] env[65788]: DEBUG nova.objects.instance [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lazy-loading 'resources' on Instance uuid 22cd2208-ebb5-401a-a06d-3dd6cd591869 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.796423] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.796783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.797074] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.797295] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.797498] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.800028] env[65788]: INFO nova.compute.manager [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Terminating instance [ 1177.893698] env[65788]: DEBUG nova.compute.manager [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1177.893897] env[65788]: DEBUG nova.compute.manager [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing instance network info cache due to event network-changed-dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1177.894743] env[65788]: DEBUG oslo_concurrency.lockutils [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Acquiring lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.894902] env[65788]: DEBUG oslo_concurrency.lockutils [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Acquired lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.895142] env[65788]: DEBUG nova.network.neutron [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Refreshing network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1177.989043] env[65788]: DEBUG nova.compute.manager [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1177.989323] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1177.990313] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50872104-db00-416a-9350-e1c5c7ac411d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.999344] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.000029] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a9c82da-5d26-4020-b876-874b97d378e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.007932] env[65788]: DEBUG oslo_vmware.api [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1178.007932] env[65788]: value = "task-4663275" [ 1178.007932] env[65788]: _type = "Task" [ 1178.007932] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.020874] env[65788]: DEBUG oslo_vmware.api [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.039917] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.041307] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.041307] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "9946cef4-b8c7-41b4-9299-eace1929bf3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.041307] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.041307] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.044670] env[65788]: INFO nova.compute.manager [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Terminating instance [ 1178.159846] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663274, 'name': CreateVM_Task, 'duration_secs': 0.560168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.160089] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1178.160861] env[65788]: WARNING neutronclient.v2_0.client [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1178.161481] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.161770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.162303] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1178.162705] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dbbb664-0345-4363-83c4-a2c1951ea3ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.170487] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1178.170487] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521997f4-fb42-89c3-b888-427671ef3ff9" [ 1178.170487] env[65788]: _type = "Task" [ 1178.170487] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.182801] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521997f4-fb42-89c3-b888-427671ef3ff9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.304517] env[65788]: DEBUG nova.compute.manager [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1178.304670] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1178.305714] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa55a12-07ef-4cac-801b-605f11b318f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.315770] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.319058] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb9ef1e4-f7bc-4b2c-b983-151bd2d0f389 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.328176] env[65788]: DEBUG oslo_vmware.api [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1178.328176] env[65788]: value = "task-4663276" [ 1178.328176] env[65788]: _type = "Task" [ 1178.328176] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.342359] env[65788]: DEBUG oslo_vmware.api [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663276, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.401187] env[65788]: WARNING neutronclient.v2_0.client [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1178.401888] env[65788]: WARNING openstack [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1178.402278] env[65788]: WARNING openstack [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1178.446782] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62a5df3-9627-4c1b-9660-a78083ad6319 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.456250] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f5d7ae-a492-430d-9c2a-1b29e535884b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.494459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b49b811-7cd7-43b9-ae44-3bc9193fc040 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.503805] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05922dec-c58d-4cae-8e52-81b655259019 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.525039] env[65788]: DEBUG nova.compute.provider_tree [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.535042] env[65788]: DEBUG oslo_vmware.api [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663275, 'name': PowerOffVM_Task, 'duration_secs': 0.277153} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.536268] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.536508] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1178.537263] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f74da870-ec60-43cc-83a8-e707106af2ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.549423] env[65788]: DEBUG nova.compute.manager [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1178.549660] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1178.551256] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e4fb91-ebdb-4525-b7ff-d83076f0e955 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.561998] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.562477] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24cb5e39-9c10-4692-818b-3332ff6fe1e7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.573038] env[65788]: DEBUG oslo_vmware.api [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1178.573038] env[65788]: value = "task-4663278" [ 1178.573038] env[65788]: _type = "Task" [ 1178.573038] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.583651] env[65788]: DEBUG oslo_vmware.api [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663278, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.632441] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1178.632441] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1178.632614] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Deleting the datastore file [datastore2] f59d609a-e5bb-4202-ac05-d486bbf9e05b {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1178.632963] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2282d652-f220-452c-a5a4-e9f32060177b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.641988] env[65788]: DEBUG oslo_vmware.api [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1178.641988] env[65788]: value = "task-4663279" [ 1178.641988] env[65788]: _type = "Task" [ 1178.641988] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.650941] env[65788]: WARNING openstack [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1178.651450] env[65788]: WARNING openstack [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1178.663693] env[65788]: DEBUG oslo_vmware.api [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.683170] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.683624] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Processing image aa3a5ddd-40cd-4c98-8384-9c40c7390261 {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.683898] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.684089] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquired lock "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.684307] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.689246] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6328112-8cf0-4b64-95f8-db0a245b4115 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.699773] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.700022] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.700991] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6a838fd-5db3-4ec8-b282-6cbb119d6d80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.708453] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1178.708453] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b6f235-7840-c94e-e55f-ddf0940bc004" [ 1178.708453] env[65788]: _type = "Task" [ 1178.708453] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.717691] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52b6f235-7840-c94e-e55f-ddf0940bc004, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.726252] env[65788]: WARNING neutronclient.v2_0.client [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1178.727116] env[65788]: WARNING openstack [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1178.727439] env[65788]: WARNING openstack [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1178.819907] env[65788]: DEBUG nova.network.neutron [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updated VIF entry in instance network info cache for port dc5727b9-f909-4ed4-80d0-06ba5986f0cc. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1178.820571] env[65788]: DEBUG nova.network.neutron [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [{"id": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "address": "fa:16:3e:5c:12:1f", "network": {"id": "0c9ee1ed-927d-41cb-b260-9a0396d20601", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-11492267-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b70d3c8627449eaa6372ebe3bd90233", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc5727b9-f9", "ovs_interfaceid": "dc5727b9-f909-4ed4-80d0-06ba5986f0cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1178.843305] env[65788]: DEBUG oslo_vmware.api [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663276, 'name': PowerOffVM_Task, 'duration_secs': 0.214128} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.843689] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.843923] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1178.844296] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2162a1e-35e3-4c93-b3fb-5173492b344f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.923881] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1178.924136] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1178.924382] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore2] 31853be1-f03d-4cd9-91d7-e9542dc1cadc {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1178.924707] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55f042c8-588c-48f2-9f4d-3223cacb68c7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.933259] env[65788]: DEBUG oslo_vmware.api [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1178.933259] env[65788]: value = "task-4663281" [ 1178.933259] env[65788]: _type = "Task" [ 1178.933259] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.942011] env[65788]: DEBUG oslo_vmware.api [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663281, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.037384] env[65788]: DEBUG nova.scheduler.client.report [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1179.083605] env[65788]: DEBUG oslo_vmware.api [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663278, 'name': PowerOffVM_Task, 'duration_secs': 0.252617} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.083967] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1179.083967] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1179.084284] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfbde291-682c-4845-abbf-6c2873b4db0d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.157458] env[65788]: DEBUG oslo_vmware.api [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233319} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.158092] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.158228] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1179.158450] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1179.158694] env[65788]: INFO nova.compute.manager [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1179.159075] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1179.159210] env[65788]: DEBUG nova.compute.manager [-] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1179.159313] env[65788]: DEBUG nova.network.neutron [-] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1179.159746] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.160299] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.160430] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.168709] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1179.169212] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1179.169212] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Deleting the datastore file [datastore2] 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1179.169532] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7181e2f0-9573-4c1c-b64a-f39dfd9062f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.177404] env[65788]: DEBUG oslo_vmware.api [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for the task: (returnval){ [ 1179.177404] env[65788]: value = "task-4663283" [ 1179.177404] env[65788]: _type = "Task" [ 1179.177404] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.188051] env[65788]: DEBUG oslo_vmware.api [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.220315] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1179.220631] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Fetch image to [datastore1] OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e/OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1179.220831] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Downloading stream optimized image aa3a5ddd-40cd-4c98-8384-9c40c7390261 to [datastore1] OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e/OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e.vmdk on the data store datastore1 as vApp {{(pid=65788) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1179.221017] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Downloading image file data aa3a5ddd-40cd-4c98-8384-9c40c7390261 to the ESX as VM named 'OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e' {{(pid=65788) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1179.252093] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.311413] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1179.311413] env[65788]: value = "resgroup-9" [ 1179.311413] env[65788]: _type = "ResourcePool" [ 1179.311413] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1179.311660] env[65788]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-2016e37f-f89a-4a1e-90b5-e3f739a32e07 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.332964] env[65788]: DEBUG oslo_concurrency.lockutils [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] Releasing lock "refresh_cache-b9d8a32e-1f0f-405c-9543-9e53046d248e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1179.333341] env[65788]: DEBUG nova.compute.manager [req-15521d9b-59e1-405d-b077-9ef8d6046ba2 req-540ce1ec-0a25-4cc3-bf02-68a370a6f8f3 service nova] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Received event network-vif-deleted-98f4edb9-9923-4aa2-9ee0-e4db751e1691 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1179.341465] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease: (returnval){ [ 1179.341465] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520848b7-5d73-b5f7-29f9-720118f88183" [ 1179.341465] env[65788]: _type = "HttpNfcLease" [ 1179.341465] env[65788]: } obtained for vApp import into resource pool (val){ [ 1179.341465] env[65788]: value = "resgroup-9" [ 1179.341465] env[65788]: _type = "ResourcePool" [ 1179.341465] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1179.341465] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the lease: (returnval){ [ 1179.341465] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520848b7-5d73-b5f7-29f9-720118f88183" [ 1179.341465] env[65788]: _type = "HttpNfcLease" [ 1179.341465] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1179.354053] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1179.354053] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520848b7-5d73-b5f7-29f9-720118f88183" [ 1179.354053] env[65788]: _type = "HttpNfcLease" [ 1179.354053] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1179.446491] env[65788]: DEBUG oslo_vmware.api [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663281, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149456} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.446786] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.446973] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1179.448067] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1179.448067] env[65788]: INFO nova.compute.manager [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1179.448067] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1179.448067] env[65788]: DEBUG nova.compute.manager [-] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1179.448067] env[65788]: DEBUG nova.network.neutron [-] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1179.448396] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.449253] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.449711] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.543556] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.795s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.573470] env[65788]: INFO nova.scheduler.client.report [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Deleted allocations for instance 22cd2208-ebb5-401a-a06d-3dd6cd591869 [ 1179.692622] env[65788]: DEBUG oslo_vmware.api [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Task: {'id': task-4663283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149088} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.693052] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.693349] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1179.693643] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1179.693949] env[65788]: INFO nova.compute.manager [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1179.694377] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1179.694678] env[65788]: DEBUG nova.compute.manager [-] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1179.694825] env[65788]: DEBUG nova.network.neutron [-] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1179.695195] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1179.696051] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1179.696536] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1179.849376] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1179.849376] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520848b7-5d73-b5f7-29f9-720118f88183" [ 1179.849376] env[65788]: _type = "HttpNfcLease" [ 1179.849376] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1179.849813] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1179.849813] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]520848b7-5d73-b5f7-29f9-720118f88183" [ 1179.849813] env[65788]: _type = "HttpNfcLease" [ 1179.849813] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1179.850499] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d64856b0-f673-4607-9b29-2629d0094501 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.860242] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520070a6-58f5-12d3-ca40-60673b19d867/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1179.860242] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520070a6-58f5-12d3-ca40-60673b19d867/disk-0.vmdk. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1179.924445] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8e5a348f-e6e5-4864-b25e-6e50c85d7276 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.998195] env[65788]: DEBUG nova.network.neutron [-] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1180.083051] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0a68aed1-4791-4ae8-8be2-3aa306de556d tempest-InstanceActionsNegativeTestJSON-1571355249 tempest-InstanceActionsNegativeTestJSON-1571355249-project-member] Lock "22cd2208-ebb5-401a-a06d-3dd6cd591869" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.934s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.298162] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1180.323212] env[65788]: DEBUG nova.compute.manager [req-44c920a5-68ea-4a81-aaff-ac3703433c13 req-64b1827c-e633-49ef-b6c7-44c6354f0de1 service nova] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Received event network-vif-deleted-458887b5-5ecd-41a3-9bb5-4ee7926f28ad {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1180.328591] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1180.501753] env[65788]: INFO nova.compute.manager [-] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Took 1.34 seconds to deallocate network for instance. [ 1181.011733] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.013028] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.013028] env[65788]: DEBUG nova.objects.instance [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lazy-loading 'resources' on Instance uuid f59d609a-e5bb-4202-ac05-d486bbf9e05b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1181.055170] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1181.055170] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520070a6-58f5-12d3-ca40-60673b19d867/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1181.056273] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4579050-661d-48c0-b0b0-3884f5661de7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.065626] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520070a6-58f5-12d3-ca40-60673b19d867/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1181.065807] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520070a6-58f5-12d3-ca40-60673b19d867/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1181.066662] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a5567fba-ff35-4f00-ab34-df0bcbd33466 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.280029] env[65788]: DEBUG oslo_vmware.rw_handles [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520070a6-58f5-12d3-ca40-60673b19d867/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1181.280029] env[65788]: INFO nova.virt.vmwareapi.images [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Downloaded image file data aa3a5ddd-40cd-4c98-8384-9c40c7390261 [ 1181.281777] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8947f35-1938-4ca6-8e47-d0bf0a6a28b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.304852] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8497ce3-ac2b-4454-a232-5958baf73e44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.334988] env[65788]: INFO nova.virt.vmwareapi.images [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] The imported VM was unregistered [ 1181.342676] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1181.343342] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Creating directory with path [datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261 {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1181.343342] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0adf03c8-f5a7-4407-ba18-5dece0eb5e9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.362665] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Created directory with path [datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261 {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1181.362981] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e/OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e.vmdk to [datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk. {{(pid=65788) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1181.363182] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-7a0f8739-e189-48b8-94bb-0c4ff6fb2f4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.371550] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1181.371550] env[65788]: value = "task-4663286" [ 1181.371550] env[65788]: _type = "Task" [ 1181.371550] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.383757] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.455613] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.455613] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.482670] env[65788]: DEBUG nova.network.neutron [-] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1181.582917] env[65788]: DEBUG nova.compute.manager [req-b84f2f58-f572-4bba-b51d-23b8eccc5a89 req-638e5677-969c-4e28-a665-4a469ee0ee6c service nova] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Received event network-vif-deleted-4c3f94f3-891e-4229-9363-1f8123b187f7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1181.738502] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796869ab-fafb-4bdc-8367-7ae2455d28ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.747650] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63c594d-11cc-421e-9502-ff424c43743d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.787041] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be5e1f5-7217-4b84-a8c2-7b3707a67b2e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.796516] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b9fb92-d5c6-421a-8b28-79c88dbd0a99 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.814794] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1181.885301] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.959465] env[65788]: DEBUG nova.compute.utils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1181.986249] env[65788]: INFO nova.compute.manager [-] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Took 2.54 seconds to deallocate network for instance. [ 1182.018649] env[65788]: DEBUG nova.network.neutron [-] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1182.348244] env[65788]: ERROR nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [req-c47a6ce8-6a16-400a-9a31-8234de7416af] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c47a6ce8-6a16-400a-9a31-8234de7416af"}]} [ 1182.369648] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1182.389014] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.391130] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1182.391233] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1182.411481] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1182.441590] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1182.463802] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.497386] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.525077] env[65788]: INFO nova.compute.manager [-] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Took 2.83 seconds to deallocate network for instance. [ 1182.645569] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0abc08-c05c-4fe2-81c9-619cf03ff7af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.658499] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011e285a-48d2-4bab-b234-596f0d990f47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.697665] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647920f4-b12b-4d2a-8420-e2dab68a4724 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.708398] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1205449-e99c-4baa-91b2-2c6ef502539a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.728490] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1182.885991] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.032527] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.254597] env[65788]: ERROR nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [req-fed0964b-6db4-4595-9941-ba45d8b4a430] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 3a6417f0-b161-4086-8a0e-1cb139eae377. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fed0964b-6db4-4595-9941-ba45d8b4a430"}]} [ 1183.277303] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1183.301650] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1183.301972] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1183.315309] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1183.338674] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1183.389063] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.394338] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "0636e867-8214-473e-9171-af66a53818a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.394533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "0636e867-8214-473e-9171-af66a53818a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.530811] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06dc1c2-02ee-4e49-8d5d-14eb06daa9b8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.539607] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95a0b6a-5e4f-45a2-a97e-3040d189e65d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.578157] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25bb4ae-bdff-4510-82e7-8e540cec078f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.581480] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.581751] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.581976] env[65788]: INFO nova.compute.manager [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Attaching volume 81b8d762-0018-4606-bbe9-c85dccaac935 to /dev/sdb [ 1183.592700] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f2696e-e1a9-4d33-ac9e-88febb08dfaf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.611515] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1183.629214] env[65788]: DEBUG nova.compute.manager [req-b17b363c-36c0-4548-9204-abade600f492 req-f3e48401-42e4-4fa4-9478-a2dc7681774d service nova] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Received event network-vif-deleted-3ace6194-8918-410f-a12e-0f966ea6e08e {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1183.637717] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b61667c-dccf-4860-bd32-96423b8fd332 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.647768] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4498d1b-9728-4c03-b4df-aa45847607e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.668774] env[65788]: DEBUG nova.virt.block_device [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updating existing volume attachment record: ed93994e-ed5a-468e-962a-9c960f6d95d8 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1183.886515] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.901031] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1184.152994] env[65788]: DEBUG nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updated inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with generation 160 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1184.153595] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 generation from 160 to 161 during operation: update_inventory {{(pid=65788) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1184.155206] env[65788]: DEBUG nova.compute.provider_tree [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1184.386497] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663286, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.731035} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.386774] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e/OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e.vmdk to [datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk. [ 1184.386973] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Cleaning up location [datastore1] OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1184.387165] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_05b18d0e-b356-4d38-a6ad-2a9d9242232e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1184.387439] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ebb46f8-f551-4262-ba21-b4c27771529e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.394654] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1184.394654] env[65788]: value = "task-4663288" [ 1184.394654] env[65788]: _type = "Task" [ 1184.394654] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.403570] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663288, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.423927] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.662669] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.650s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.665236] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.171s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.665575] env[65788]: DEBUG nova.objects.instance [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid 31853be1-f03d-4cd9-91d7-e9542dc1cadc {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.686558] env[65788]: INFO nova.scheduler.client.report [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Deleted allocations for instance f59d609a-e5bb-4202-ac05-d486bbf9e05b [ 1184.904680] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663288, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042794} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.904920] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.905063] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Releasing lock "[datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1184.905315] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk to [datastore1] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1184.905611] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f883625-710c-430c-a528-2abab9a1ed81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.914905] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1184.914905] env[65788]: value = "task-4663289" [ 1184.914905] env[65788]: _type = "Task" [ 1184.914905] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.923666] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663289, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.197148] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b85bca84-2d37-4f90-9361-2db3e44b8f3d tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "f59d609a-e5bb-4202-ac05-d486bbf9e05b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.717s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1185.332607] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54d005e-32c1-415a-83a2-5cda5d88d09a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.341712] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ad224f-e754-44d9-9a0b-e99e89f1aac8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.376312] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bc949d-483b-4538-a9e1-3f1463eb2fa4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.385444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed66f2c0-fc53-4161-9216-2243a53324dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.401808] env[65788]: DEBUG nova.compute.provider_tree [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1185.429197] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663289, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.906100] env[65788]: DEBUG nova.scheduler.client.report [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1185.909732] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "6012182d-8b4e-4379-949f-d3a26211699b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1185.909974] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "6012182d-8b4e-4379-949f-d3a26211699b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1185.910226] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "6012182d-8b4e-4379-949f-d3a26211699b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1185.910417] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "6012182d-8b4e-4379-949f-d3a26211699b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1185.910572] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "6012182d-8b4e-4379-949f-d3a26211699b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1185.912548] env[65788]: INFO nova.compute.manager [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Terminating instance [ 1185.930167] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663289, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.414195] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.417123] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.385s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.417382] env[65788]: DEBUG nova.objects.instance [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lazy-loading 'resources' on Instance uuid 9946cef4-b8c7-41b4-9299-eace1929bf3d {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.418972] env[65788]: DEBUG nova.compute.manager [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1186.419233] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1186.420442] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb89e4a-b629-4089-b6f6-7933074b458f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.434531] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663289, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.436954] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1186.437264] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd55e9ad-8775-48d7-afc1-974c93c7e95f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.440179] env[65788]: INFO nova.scheduler.client.report [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance 31853be1-f03d-4cd9-91d7-e9542dc1cadc [ 1186.447485] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1186.447485] env[65788]: value = "task-4663291" [ 1186.447485] env[65788]: _type = "Task" [ 1186.447485] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.457385] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663291, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.934082] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663289, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.948412] env[65788]: DEBUG oslo_concurrency.lockutils [None req-22bf1762-de74-4d7b-82b0-d07a96f8372c tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "31853be1-f03d-4cd9-91d7-e9542dc1cadc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.152s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.964055] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663291, 'name': PowerOffVM_Task, 'duration_secs': 0.250873} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.964450] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.964673] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1186.965013] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e942bee-a487-40ee-84c7-4b30206f3a38 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.051082] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.051469] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.051742] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Deleting the datastore file [datastore1] 6012182d-8b4e-4379-949f-d3a26211699b {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.052123] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccadd31a-9ac1-4880-b758-ba16f7a5719e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.060751] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for the task: (returnval){ [ 1187.060751] env[65788]: value = "task-4663293" [ 1187.060751] env[65788]: _type = "Task" [ 1187.060751] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.066162] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0892e993-9449-4d33-911d-38a5fedb2d41 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.076660] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663293, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.080142] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44e3621-1db0-4bf5-aea5-174d2d03f45c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.116875] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4673641-7fe2-4941-8642-23daf3a3f620 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.125820] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c21e76-e298-403b-a500-32220a17b93e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.142177] env[65788]: DEBUG nova.compute.provider_tree [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.431233] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663289, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.41987} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.431513] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/aa3a5ddd-40cd-4c98-8384-9c40c7390261/aa3a5ddd-40cd-4c98-8384-9c40c7390261.vmdk to [datastore1] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1187.432375] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5558728f-52f7-4039-8cdf-28b22b753491 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.455430] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1187.455747] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0cb1dd3-8800-43f2-8e17-efe7e491e78f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.476991] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1187.476991] env[65788]: value = "task-4663294" [ 1187.476991] env[65788]: _type = "Task" [ 1187.476991] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.486137] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663294, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.572636] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663293, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.645434] env[65788]: DEBUG nova.scheduler.client.report [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1187.707273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "d15953f8-f74a-422a-867d-fa835bb479ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.707551] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "d15953f8-f74a-422a-867d-fa835bb479ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.988685] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663294, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.071840] env[65788]: DEBUG oslo_vmware.api [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Task: {'id': task-4663293, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.594517} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.072169] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.072419] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.072621] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.072802] env[65788]: INFO nova.compute.manager [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Took 1.65 seconds to destroy the instance on the hypervisor. [ 1188.073064] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1188.073268] env[65788]: DEBUG nova.compute.manager [-] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1188.073367] env[65788]: DEBUG nova.network.neutron [-] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1188.073682] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1188.074298] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1188.074728] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1188.151580] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.154316] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.730s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1188.156228] env[65788]: INFO nova.compute.claims [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1188.187821] env[65788]: INFO nova.scheduler.client.report [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Deleted allocations for instance 9946cef4-b8c7-41b4-9299-eace1929bf3d [ 1188.210967] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1188.224604] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1188.224873] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910436', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'name': 'volume-81b8d762-0018-4606-bbe9-c85dccaac935', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '50c70e36-a3dc-4e8c-8fe6-74039d1828bf', 'attached_at': '', 'detached_at': '', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'serial': '81b8d762-0018-4606-bbe9-c85dccaac935'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1188.226255] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe938377-a068-45ad-8489-7f5faf09f103 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.233333] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1188.247706] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d430e57-af00-452f-8e70-84f7dc789fa8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.275024] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] volume-81b8d762-0018-4606-bbe9-c85dccaac935/volume-81b8d762-0018-4606-bbe9-c85dccaac935.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.275390] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a9f607e-190d-45e2-886f-d6ba4a2092e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.295175] env[65788]: DEBUG oslo_vmware.api [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1188.295175] env[65788]: value = "task-4663295" [ 1188.295175] env[65788]: _type = "Task" [ 1188.295175] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.305875] env[65788]: DEBUG oslo_vmware.api [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.489471] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663294, 'name': ReconfigVM_Task, 'duration_secs': 0.812829} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.491372] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Reconfigured VM instance instance-00000067 to attach disk [datastore1] b9d8a32e-1f0f-405c-9543-9e53046d248e/b9d8a32e-1f0f-405c-9543-9e53046d248e.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.492468] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d31658a6-3b8a-41c9-b0a7-dd8a5d813850 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.503068] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1188.503068] env[65788]: value = "task-4663296" [ 1188.503068] env[65788]: _type = "Task" [ 1188.503068] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.513056] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663296, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.528858] env[65788]: DEBUG nova.compute.manager [req-8f1f9267-1369-4a3d-9887-9fa21accce94 req-c4512ca9-d5b7-4a7d-ae9c-84271fe5bdbc service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Received event network-vif-deleted-dfa97bd4-874e-4c84-a6d1-d7072b745f83 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1188.529279] env[65788]: INFO nova.compute.manager [req-8f1f9267-1369-4a3d-9887-9fa21accce94 req-c4512ca9-d5b7-4a7d-ae9c-84271fe5bdbc service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Neutron deleted interface dfa97bd4-874e-4c84-a6d1-d7072b745f83; detaching it from the instance and deleting it from the info cache [ 1188.529530] env[65788]: DEBUG nova.network.neutron [req-8f1f9267-1369-4a3d-9887-9fa21accce94 req-c4512ca9-d5b7-4a7d-ae9c-84271fe5bdbc service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1188.697520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-28bc5844-8dff-4e5c-892c-8aeb2c1d2949 tempest-AttachVolumeTestJSON-999763968 tempest-AttachVolumeTestJSON-999763968-project-member] Lock "9946cef4-b8c7-41b4-9299-eace1929bf3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.657s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.735788] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.807238] env[65788]: DEBUG oslo_vmware.api [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663295, 'name': ReconfigVM_Task, 'duration_secs': 0.437476} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.807571] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Reconfigured VM instance instance-0000006d to attach disk [datastore2] volume-81b8d762-0018-4606-bbe9-c85dccaac935/volume-81b8d762-0018-4606-bbe9-c85dccaac935.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.813226] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e093c058-7938-456e-aab9-3b85812c16b5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.832138] env[65788]: DEBUG oslo_vmware.api [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1188.832138] env[65788]: value = "task-4663297" [ 1188.832138] env[65788]: _type = "Task" [ 1188.832138] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.841489] env[65788]: DEBUG oslo_vmware.api [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663297, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.978935] env[65788]: DEBUG nova.network.neutron [-] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1189.014608] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663296, 'name': Rename_Task, 'duration_secs': 0.232266} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.014948] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.015242] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8585bf5e-91df-4b94-aadf-684c511b1727 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.022387] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1189.022387] env[65788]: value = "task-4663299" [ 1189.022387] env[65788]: _type = "Task" [ 1189.022387] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.032476] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.035331] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c5f4242-72d3-43aa-a1a1-52f8f3279505 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.046119] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c31d40-30cf-4444-926b-aca1988e2731 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.079389] env[65788]: DEBUG nova.compute.manager [req-8f1f9267-1369-4a3d-9887-9fa21accce94 req-c4512ca9-d5b7-4a7d-ae9c-84271fe5bdbc service nova] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Detach interface failed, port_id=dfa97bd4-874e-4c84-a6d1-d7072b745f83, reason: Instance 6012182d-8b4e-4379-949f-d3a26211699b could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1189.327445] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c930ad5e-8962-4aa7-877c-70d32ee0aeca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.340199] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cfebbaf-19a3-44a8-893e-2fc2c8964492 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.347103] env[65788]: DEBUG oslo_vmware.api [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663297, 'name': ReconfigVM_Task, 'duration_secs': 0.153294} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.347840] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910436', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'name': 'volume-81b8d762-0018-4606-bbe9-c85dccaac935', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '50c70e36-a3dc-4e8c-8fe6-74039d1828bf', 'attached_at': '', 'detached_at': '', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'serial': '81b8d762-0018-4606-bbe9-c85dccaac935'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1189.376232] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709168e5-7ff6-4b8f-b468-5d7a82fac328 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.387618] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca5383b-b059-458c-91cf-4170b759b0c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.402063] env[65788]: DEBUG nova.compute.provider_tree [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.481824] env[65788]: INFO nova.compute.manager [-] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Took 1.41 seconds to deallocate network for instance. [ 1189.534653] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663299, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.905045] env[65788]: DEBUG nova.scheduler.client.report [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1189.988850] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.034201] env[65788]: DEBUG oslo_vmware.api [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663299, 'name': PowerOnVM_Task, 'duration_secs': 0.685395} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.034544] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1190.138119] env[65788]: DEBUG nova.compute.manager [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1190.139051] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464f0794-6076-4d4a-a493-def900fa1bff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.199500] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.199798] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.392067] env[65788]: DEBUG nova.objects.instance [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'flavor' on Instance uuid 50c70e36-a3dc-4e8c-8fe6-74039d1828bf {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1190.411122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.411706] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1190.414476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.679s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.415949] env[65788]: INFO nova.compute.claims [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1190.638397] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.656742] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b39d2536-44b5-4d1c-a7d3-d381db03641c tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.425s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.657837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.259s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.657837] env[65788]: INFO nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] During sync_power_state the instance has a pending task (spawning). Skip. [ 1190.657964] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.902040] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b344e45c-31d9-40ff-9808-e18dba4f47ab tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.320s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1190.920524] env[65788]: DEBUG nova.compute.utils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1190.923867] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1190.924079] env[65788]: DEBUG nova.network.neutron [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1190.924458] env[65788]: WARNING neutronclient.v2_0.client [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1190.924791] env[65788]: WARNING neutronclient.v2_0.client [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1190.925489] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1190.925836] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1190.938957] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1191.005381] env[65788]: DEBUG nova.policy [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6a3f4e7a11a740738485f0f81c1bbd0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3326bb2dfdd54781b984431df7ae4609', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1191.086952] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e422bb24-9f4a-4047-8999-9790611df117 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.096049] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1823437-dc9e-461c-a593-92b3ae0b1711 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.129623] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bffe119-7e0e-42fd-b85c-bfe3e65ffc90 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.139632] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b7af1d-526c-4001-88ca-facc46f89bcd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.144648] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.144648] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.160546] env[65788]: DEBUG nova.compute.provider_tree [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1191.347257] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1191.347705] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1191.426675] env[65788]: DEBUG nova.network.neutron [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Successfully created port: b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1191.640189] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.649259] env[65788]: INFO nova.compute.manager [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Detaching volume 81b8d762-0018-4606-bbe9-c85dccaac935 [ 1191.666654] env[65788]: DEBUG nova.scheduler.client.report [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1191.694961] env[65788]: INFO nova.virt.block_device [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Attempting to driver detach volume 81b8d762-0018-4606-bbe9-c85dccaac935 from mountpoint /dev/sdb [ 1191.695259] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1191.695516] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910436', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'name': 'volume-81b8d762-0018-4606-bbe9-c85dccaac935', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '50c70e36-a3dc-4e8c-8fe6-74039d1828bf', 'attached_at': '', 'detached_at': '', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'serial': '81b8d762-0018-4606-bbe9-c85dccaac935'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1191.696504] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002a6b5a-dc25-4383-a8cb-9d8d4676c70e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.731163] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34237761-237b-4398-aefa-8eb2e9152de3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.742294] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d37fe96-88cb-4ce9-bcf0-b943512323a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.764218] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.764529] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.764821] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.765015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.765193] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.767523] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd64b706-9d84-4444-936d-1274ec412528 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.770766] env[65788]: INFO nova.compute.manager [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Terminating instance [ 1191.786470] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The volume has not been displaced from its original location: [datastore2] volume-81b8d762-0018-4606-bbe9-c85dccaac935/volume-81b8d762-0018-4606-bbe9-c85dccaac935.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1191.791778] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1191.792387] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f815237e-cf7a-40aa-b410-765d993a51d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.812733] env[65788]: DEBUG oslo_vmware.api [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1191.812733] env[65788]: value = "task-4663300" [ 1191.812733] env[65788]: _type = "Task" [ 1191.812733] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.822350] env[65788]: DEBUG oslo_vmware.api [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663300, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.947987] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1191.980199] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1191.980459] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1191.980616] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1191.980798] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1191.980944] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1191.981393] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1191.981685] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.981890] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1191.982122] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1191.982405] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1191.982729] env[65788]: DEBUG nova.virt.hardware [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1191.984155] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa5a59f-3390-483f-97ca-d06073167103 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.994696] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44400ed9-60c0-4046-a736-1c656e2d109f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.172832] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.757s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.172832] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1192.175756] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.187s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1192.175892] env[65788]: DEBUG nova.objects.instance [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lazy-loading 'resources' on Instance uuid 6012182d-8b4e-4379-949f-d3a26211699b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.274959] env[65788]: DEBUG nova.compute.manager [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1192.275364] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1192.276310] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c295cb2a-76f7-42af-ac03-42c30ba20653 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.285252] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1192.285549] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99ec3606-8fc0-4bf6-ab44-37bd92424e57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.293023] env[65788]: DEBUG oslo_vmware.api [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1192.293023] env[65788]: value = "task-4663302" [ 1192.293023] env[65788]: _type = "Task" [ 1192.293023] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.304557] env[65788]: DEBUG oslo_vmware.api [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663302, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.323540] env[65788]: DEBUG oslo_vmware.api [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663300, 'name': ReconfigVM_Task, 'duration_secs': 0.237527} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.323882] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1192.328925] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b57b670-7d53-4531-9435-d803e344e86d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.346017] env[65788]: DEBUG oslo_vmware.api [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1192.346017] env[65788]: value = "task-4663303" [ 1192.346017] env[65788]: _type = "Task" [ 1192.346017] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.357083] env[65788]: DEBUG oslo_vmware.api [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663303, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.638749] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1192.679307] env[65788]: DEBUG nova.compute.utils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1192.684874] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1192.685302] env[65788]: DEBUG nova.network.neutron [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1192.685588] env[65788]: WARNING neutronclient.v2_0.client [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.685910] env[65788]: WARNING neutronclient.v2_0.client [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1192.686516] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1192.686858] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1192.746096] env[65788]: DEBUG nova.policy [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1192.803991] env[65788]: DEBUG oslo_vmware.api [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663302, 'name': PowerOffVM_Task, 'duration_secs': 0.206889} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.806792] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1192.806978] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1192.807468] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ba92e34-2fa0-4dff-9fc4-c3d18f34317d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.834426] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a1997b-88b9-4a75-b1a6-797fa590d528 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.843313] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03537e3-4e41-4e06-ae01-2d5fcdd620ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.856392] env[65788]: DEBUG oslo_vmware.api [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663303, 'name': ReconfigVM_Task, 'duration_secs': 0.187485} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.885242] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910436', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'name': 'volume-81b8d762-0018-4606-bbe9-c85dccaac935', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '50c70e36-a3dc-4e8c-8fe6-74039d1828bf', 'attached_at': '', 'detached_at': '', 'volume_id': '81b8d762-0018-4606-bbe9-c85dccaac935', 'serial': '81b8d762-0018-4606-bbe9-c85dccaac935'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1192.891877] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212f813a-bba8-46bc-acd9-2cbb9aa7df20 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.895401] env[65788]: DEBUG nova.compute.manager [req-ea46bd22-4a4b-4527-914a-2c3c3b97713a req-13cf6d1f-4857-4cc6-818b-668bacb06a64 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Received event network-vif-plugged-b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1192.895620] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea46bd22-4a4b-4527-914a-2c3c3b97713a req-13cf6d1f-4857-4cc6-818b-668bacb06a64 service nova] Acquiring lock "0636e867-8214-473e-9171-af66a53818a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1192.895857] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea46bd22-4a4b-4527-914a-2c3c3b97713a req-13cf6d1f-4857-4cc6-818b-668bacb06a64 service nova] Lock "0636e867-8214-473e-9171-af66a53818a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1192.896047] env[65788]: DEBUG oslo_concurrency.lockutils [req-ea46bd22-4a4b-4527-914a-2c3c3b97713a req-13cf6d1f-4857-4cc6-818b-668bacb06a64 service nova] Lock "0636e867-8214-473e-9171-af66a53818a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1192.896240] env[65788]: DEBUG nova.compute.manager [req-ea46bd22-4a4b-4527-914a-2c3c3b97713a req-13cf6d1f-4857-4cc6-818b-668bacb06a64 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] No waiting events found dispatching network-vif-plugged-b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1192.896403] env[65788]: WARNING nova.compute.manager [req-ea46bd22-4a4b-4527-914a-2c3c3b97713a req-13cf6d1f-4857-4cc6-818b-668bacb06a64 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Received unexpected event network-vif-plugged-b7dbcca3-428b-4b01-9903-f12350d7be4d for instance with vm_state building and task_state spawning. [ 1192.897212] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1192.897367] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1192.897544] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleting the datastore file [datastore1] b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1192.897817] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c0f8dd7-79fd-49ec-913d-eda417dd9d25 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.906582] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bcf06d-0c52-41d3-a999-d8caacdd9bd2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.915032] env[65788]: DEBUG oslo_vmware.api [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for the task: (returnval){ [ 1192.915032] env[65788]: value = "task-4663305" [ 1192.915032] env[65788]: _type = "Task" [ 1192.915032] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.925050] env[65788]: DEBUG nova.compute.provider_tree [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.932815] env[65788]: DEBUG oslo_vmware.api [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663305, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.969931] env[65788]: DEBUG nova.network.neutron [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Successfully updated port: b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1193.052820] env[65788]: DEBUG nova.network.neutron [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Successfully created port: 7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1193.186097] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1193.423519] env[65788]: DEBUG oslo_vmware.api [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Task: {'id': task-4663305, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158887} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.423830] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1193.423974] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1193.424199] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1193.424501] env[65788]: INFO nova.compute.manager [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1193.424770] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1193.424973] env[65788]: DEBUG nova.compute.manager [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1193.425082] env[65788]: DEBUG nova.network.neutron [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1193.425375] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.425934] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.426209] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.434225] env[65788]: DEBUG nova.scheduler.client.report [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1193.440867] env[65788]: DEBUG nova.objects.instance [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'flavor' on Instance uuid 50c70e36-a3dc-4e8c-8fe6-74039d1828bf {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.473500] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "refresh_cache-0636e867-8214-473e-9171-af66a53818a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.473500] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquired lock "refresh_cache-0636e867-8214-473e-9171-af66a53818a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1193.473717] env[65788]: DEBUG nova.network.neutron [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1193.475874] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1193.632378] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.639047] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1193.639556] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1193.941347] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.763s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.980337] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1193.980337] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1193.988434] env[65788]: INFO nova.scheduler.client.report [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Deleted allocations for instance 6012182d-8b4e-4379-949f-d3a26211699b [ 1194.035325] env[65788]: DEBUG nova.network.neutron [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1194.048613] env[65788]: DEBUG nova.compute.manager [req-7b991e98-d8c9-447e-bec6-45d35cadfb93 req-5f7d325e-5691-4d41-8c7c-bfbd9bb3453d service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Received event network-vif-deleted-dc5727b9-f909-4ed4-80d0-06ba5986f0cc {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1194.049013] env[65788]: INFO nova.compute.manager [req-7b991e98-d8c9-447e-bec6-45d35cadfb93 req-5f7d325e-5691-4d41-8c7c-bfbd9bb3453d service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Neutron deleted interface dc5727b9-f909-4ed4-80d0-06ba5986f0cc; detaching it from the instance and deleting it from the info cache [ 1194.049801] env[65788]: DEBUG nova.network.neutron [req-7b991e98-d8c9-447e-bec6-45d35cadfb93 req-5f7d325e-5691-4d41-8c7c-bfbd9bb3453d service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.060889] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.061476] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.161069] env[65788]: WARNING neutronclient.v2_0.client [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1194.161069] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1194.161577] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1194.195647] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1194.239360] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1194.240473] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1194.240750] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1194.240926] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1194.241157] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1194.241332] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1194.241584] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1194.241757] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1194.241947] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1194.242133] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1194.242410] env[65788]: DEBUG nova.virt.hardware [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1194.243488] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cf04f1-3bac-4097-ae46-39f782253ad3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.255599] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a842063-e23f-4a4c-ad44-6161b2654dbd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.403859] env[65788]: DEBUG nova.network.neutron [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.451211] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31d52071-37e1-41fa-b12e-c26fd42277c3 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.306s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.497159] env[65788]: DEBUG oslo_concurrency.lockutils [None req-61e2ea5b-9024-4e4c-bf7e-ff3bd5349cd8 tempest-ServerRescueTestJSON-1826206369 tempest-ServerRescueTestJSON-1826206369-project-member] Lock "6012182d-8b4e-4379-949f-d3a26211699b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.587s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.554286] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f8ad7f9f-818b-4ea7-bb64-ed28f3b58bb8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.565061] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce31681-873f-4b42-bca8-cabeae0b3795 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.582126] env[65788]: DEBUG nova.network.neutron [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Successfully updated port: 7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1194.603834] env[65788]: DEBUG nova.compute.manager [req-7b991e98-d8c9-447e-bec6-45d35cadfb93 req-5f7d325e-5691-4d41-8c7c-bfbd9bb3453d service nova] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Detach interface failed, port_id=dc5727b9-f909-4ed4-80d0-06ba5986f0cc, reason: Instance b9d8a32e-1f0f-405c-9543-9e53046d248e could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1194.617878] env[65788]: DEBUG nova.network.neutron [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Updating instance_info_cache with network_info: [{"id": "b7dbcca3-428b-4b01-9903-f12350d7be4d", "address": "fa:16:3e:fe:e5:5a", "network": {"id": "36cc96b0-5023-4df6-8c6a-11a9716a4166", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-829587981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3326bb2dfdd54781b984431df7ae4609", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7dbcca3-42", "ovs_interfaceid": "b7dbcca3-428b-4b01-9903-f12350d7be4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1194.619470] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1194.619556] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.619768] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1194.619960] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.620196] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.624856] env[65788]: INFO nova.compute.manager [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Terminating instance [ 1194.910107] env[65788]: INFO nova.compute.manager [-] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Took 1.48 seconds to deallocate network for instance. [ 1194.916114] env[65788]: DEBUG nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Received event network-changed-b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1194.916114] env[65788]: DEBUG nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Refreshing instance network info cache due to event network-changed-b7dbcca3-428b-4b01-9903-f12350d7be4d. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1194.916212] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Acquiring lock "refresh_cache-0636e867-8214-473e-9171-af66a53818a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.088445] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-d15953f8-f74a-422a-867d-fa835bb479ff" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.088716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-d15953f8-f74a-422a-867d-fa835bb479ff" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.088810] env[65788]: DEBUG nova.network.neutron [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1195.125753] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Releasing lock "refresh_cache-0636e867-8214-473e-9171-af66a53818a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1195.126168] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Instance network_info: |[{"id": "b7dbcca3-428b-4b01-9903-f12350d7be4d", "address": "fa:16:3e:fe:e5:5a", "network": {"id": "36cc96b0-5023-4df6-8c6a-11a9716a4166", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-829587981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3326bb2dfdd54781b984431df7ae4609", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7dbcca3-42", "ovs_interfaceid": "b7dbcca3-428b-4b01-9903-f12350d7be4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1195.126556] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Acquired lock "refresh_cache-0636e867-8214-473e-9171-af66a53818a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.126680] env[65788]: DEBUG nova.network.neutron [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Refreshing network info cache for port b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1195.127996] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:e5:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c842425c-544e-4ce2-9657-512723bd318e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7dbcca3-428b-4b01-9903-f12350d7be4d', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1195.136735] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Creating folder: Project (3326bb2dfdd54781b984431df7ae4609). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1195.139335] env[65788]: WARNING neutronclient.v2_0.client [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1195.139335] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.139335] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.146661] env[65788]: DEBUG nova.compute.manager [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1195.146908] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1195.147310] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0eeaab99-149f-4750-8fb9-f94f0e8b5880 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.150186] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2009c770-e9ae-453f-9201-7d90deda202e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.160224] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1195.160527] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa66c021-24bf-4572-b4e0-d5403d553c00 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.166924] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Created folder: Project (3326bb2dfdd54781b984431df7ae4609) in parent group-v910111. [ 1195.167175] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Creating folder: Instances. Parent ref: group-v910437. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1195.169366] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c2c4316-e90e-4b00-a634-c33f2d0821e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.171664] env[65788]: DEBUG oslo_vmware.api [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1195.171664] env[65788]: value = "task-4663307" [ 1195.171664] env[65788]: _type = "Task" [ 1195.171664] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.183257] env[65788]: DEBUG oslo_vmware.api [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663307, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.186991] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Created folder: Instances in parent group-v910437. [ 1195.187482] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1195.187839] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0636e867-8214-473e-9171-af66a53818a7] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1195.192231] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59cfa2f7-e551-485b-b168-e340812ed300 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.218271] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1195.218271] env[65788]: value = "task-4663309" [ 1195.218271] env[65788]: _type = "Task" [ 1195.218271] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.229116] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663309, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.343499] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.344133] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.420544] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.420872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.421460] env[65788]: DEBUG nova.objects.instance [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lazy-loading 'resources' on Instance uuid b9d8a32e-1f0f-405c-9543-9e53046d248e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1195.437380] env[65788]: WARNING neutronclient.v2_0.client [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1195.438081] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.438994] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.580324] env[65788]: DEBUG nova.network.neutron [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Updated VIF entry in instance network info cache for port b7dbcca3-428b-4b01-9903-f12350d7be4d. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1195.580849] env[65788]: DEBUG nova.network.neutron [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Updating instance_info_cache with network_info: [{"id": "b7dbcca3-428b-4b01-9903-f12350d7be4d", "address": "fa:16:3e:fe:e5:5a", "network": {"id": "36cc96b0-5023-4df6-8c6a-11a9716a4166", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-829587981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3326bb2dfdd54781b984431df7ae4609", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c842425c-544e-4ce2-9657-512723bd318e", "external-id": "nsx-vlan-transportzone-80", "segmentation_id": 80, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7dbcca3-42", "ovs_interfaceid": "b7dbcca3-428b-4b01-9903-f12350d7be4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1195.592288] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.592945] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.635718] env[65788]: DEBUG nova.network.neutron [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1195.638846] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.664208] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.664748] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.685794] env[65788]: DEBUG oslo_vmware.api [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663307, 'name': PowerOffVM_Task, 'duration_secs': 0.236798} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.685794] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1195.685794] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1195.685794] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b85a6263-5cbf-442b-8604-44c61c3a9d04 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.729173] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663309, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.761156] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1195.761391] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1195.761988] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleting the datastore file [datastore2] 50c70e36-a3dc-4e8c-8fe6-74039d1828bf {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1195.761988] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec2900b5-94f1-4613-a15f-6682b0fd690c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.770644] env[65788]: DEBUG oslo_vmware.api [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1195.770644] env[65788]: value = "task-4663312" [ 1195.770644] env[65788]: _type = "Task" [ 1195.770644] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.780998] env[65788]: DEBUG oslo_vmware.api [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.790458] env[65788]: WARNING neutronclient.v2_0.client [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1195.791267] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1195.791521] env[65788]: WARNING openstack [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1195.880247] env[65788]: DEBUG nova.network.neutron [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Updating instance_info_cache with network_info: [{"id": "7bd8a945-5b5b-43be-8383-f403522245b3", "address": "fa:16:3e:b9:22:8d", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bd8a945-5b", "ovs_interfaceid": "7bd8a945-5b5b-43be-8383-f403522245b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1195.905859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.906143] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.906335] env[65788]: INFO nova.compute.manager [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Shelving [ 1196.056662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f477ca-2c29-4cf0-b8f7-c8524b3560f2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.067507] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba468e1-c3dc-420d-93e4-267a1e21f970 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.099195] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Releasing lock "refresh_cache-0636e867-8214-473e-9171-af66a53818a7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.099502] env[65788]: DEBUG nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Received event network-vif-plugged-7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1196.099671] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Acquiring lock "d15953f8-f74a-422a-867d-fa835bb479ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1196.099881] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Lock "d15953f8-f74a-422a-867d-fa835bb479ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1196.100056] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Lock "d15953f8-f74a-422a-867d-fa835bb479ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1196.100225] env[65788]: DEBUG nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] No waiting events found dispatching network-vif-plugged-7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1196.100450] env[65788]: WARNING nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Received unexpected event network-vif-plugged-7bd8a945-5b5b-43be-8383-f403522245b3 for instance with vm_state building and task_state spawning. [ 1196.100589] env[65788]: DEBUG nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Received event network-changed-7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1196.100753] env[65788]: DEBUG nova.compute.manager [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Refreshing instance network info cache due to event network-changed-7bd8a945-5b5b-43be-8383-f403522245b3. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1196.100923] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Acquiring lock "refresh_cache-d15953f8-f74a-422a-867d-fa835bb479ff" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.102065] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c782cb-c661-467c-8c84-26e0e64dae77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.110916] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3272d3-4cec-4af3-81df-24a996413327 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.126727] env[65788]: DEBUG nova.compute.provider_tree [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1196.142507] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1196.230747] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663309, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.282470] env[65788]: DEBUG oslo_vmware.api [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150821} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.282941] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.283178] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1196.283373] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1196.283579] env[65788]: INFO nova.compute.manager [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1196.283863] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1196.284089] env[65788]: DEBUG nova.compute.manager [-] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1196.284191] env[65788]: DEBUG nova.network.neutron [-] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1196.284564] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.285253] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.285611] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.334516] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.383512] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-d15953f8-f74a-422a-867d-fa835bb479ff" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.384036] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Instance network_info: |[{"id": "7bd8a945-5b5b-43be-8383-f403522245b3", "address": "fa:16:3e:b9:22:8d", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bd8a945-5b", "ovs_interfaceid": "7bd8a945-5b5b-43be-8383-f403522245b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1196.384493] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Acquired lock "refresh_cache-d15953f8-f74a-422a-867d-fa835bb479ff" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.384718] env[65788]: DEBUG nova.network.neutron [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Refreshing network info cache for port 7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1196.386025] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:22:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7bd8a945-5b5b-43be-8383-f403522245b3', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1196.395883] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1196.397015] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1196.397272] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3fe9df5-8070-45ae-860c-fc51f6e18e17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.420865] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1196.420865] env[65788]: value = "task-4663313" [ 1196.420865] env[65788]: _type = "Task" [ 1196.420865] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.430490] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663313, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.629098] env[65788]: DEBUG nova.scheduler.client.report [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1196.731104] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663309, 'name': CreateVM_Task, 'duration_secs': 1.42023} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.731305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0636e867-8214-473e-9171-af66a53818a7] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1196.731820] env[65788]: WARNING neutronclient.v2_0.client [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.732599] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.732948] env[65788]: WARNING openstack [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.814706] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.815128] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.815381] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1196.816333] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43e7fe12-bc44-4a95-adb9-b8b3c6199d51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.824964] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1196.824964] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5215abc1-7a50-e28f-06ca-acaebc667c10" [ 1196.824964] env[65788]: _type = "Task" [ 1196.824964] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.833580] env[65788]: DEBUG nova.compute.manager [req-43c1cd0c-5971-4c5c-aa38-8630bbe32e06 req-c65e146b-02e5-423d-9a30-3ab72a46ef26 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Received event network-vif-deleted-731c93dc-1d06-4943-bac0-fc800f9f449d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1196.833761] env[65788]: INFO nova.compute.manager [req-43c1cd0c-5971-4c5c-aa38-8630bbe32e06 req-c65e146b-02e5-423d-9a30-3ab72a46ef26 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Neutron deleted interface 731c93dc-1d06-4943-bac0-fc800f9f449d; detaching it from the instance and deleting it from the info cache [ 1196.833927] env[65788]: DEBUG nova.network.neutron [req-43c1cd0c-5971-4c5c-aa38-8630bbe32e06 req-c65e146b-02e5-423d-9a30-3ab72a46ef26 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1196.839131] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5215abc1-7a50-e28f-06ca-acaebc667c10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.897439] env[65788]: WARNING neutronclient.v2_0.client [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.898204] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1196.898565] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1196.920074] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1196.921071] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82f1747f-79af-4243-ba31-8c38dc3f17da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.933120] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663313, 'name': CreateVM_Task, 'duration_secs': 0.451997} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.934299] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1196.934800] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1196.934800] env[65788]: value = "task-4663314" [ 1196.934800] env[65788]: _type = "Task" [ 1196.934800] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.935360] env[65788]: WARNING neutronclient.v2_0.client [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1196.935709] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.951403] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663314, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.011362] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.011746] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.083102] env[65788]: WARNING neutronclient.v2_0.client [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1197.083806] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1197.085049] env[65788]: WARNING openstack [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1197.139614] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.718s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.142450] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1197.142674] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.143034] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1197.143724] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a697d08a-79c0-4c18-b9ad-0f0fa33ff610 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.153176] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b9ace0-eedf-4d16-bcf5-7d4dcdde2cea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.170592] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b185a693-b6e4-42ec-8633-a798acf68faa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.175535] env[65788]: INFO nova.scheduler.client.report [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Deleted allocations for instance b9d8a32e-1f0f-405c-9543-9e53046d248e [ 1197.182229] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe53d91f-97ce-4e28-abf8-3a2720fbe25d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.191416] env[65788]: DEBUG nova.network.neutron [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Updated VIF entry in instance network info cache for port 7bd8a945-5b5b-43be-8383-f403522245b3. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1197.191824] env[65788]: DEBUG nova.network.neutron [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Updating instance_info_cache with network_info: [{"id": "7bd8a945-5b5b-43be-8383-f403522245b3", "address": "fa:16:3e:b9:22:8d", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7bd8a945-5b", "ovs_interfaceid": "7bd8a945-5b5b-43be-8383-f403522245b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.221850] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178541MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1197.222063] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1197.222298] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1197.225108] env[65788]: DEBUG oslo_concurrency.lockutils [req-0722c4e8-259a-4312-a6ef-acc5245fe0eb req-293f19cc-8c0e-4d06-9c92-1b7fc4731909 service nova] Releasing lock "refresh_cache-d15953f8-f74a-422a-867d-fa835bb479ff" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.321091] env[65788]: DEBUG nova.network.neutron [-] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1197.337871] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5215abc1-7a50-e28f-06ca-acaebc667c10, 'name': SearchDatastore_Task, 'duration_secs': 0.01226} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.338208] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.339920] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1197.339920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.339920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1197.339920] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1197.339920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1197.339920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1197.339920] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6658c383-1a05-4f24-886c-7e5c8b9903df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.342072] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aae90ede-483f-458f-9cac-4506cf97a8aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.345651] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7670e8f7-13a3-4ea8-84dd-5eff9f5d268d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.352743] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1197.352743] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a16517-633f-b726-34af-d815dec087d6" [ 1197.352743] env[65788]: _type = "Task" [ 1197.352743] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.363484] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec9bb64-2c4f-466e-ab3c-e4bff95ff74e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.374922] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1197.375184] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1197.376609] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eaf1df0-3c98-4c05-9266-2cd1fa2e0a2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.383976] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1197.383976] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c90236-bb6a-0a7e-4f77-9ac20620543b" [ 1197.383976] env[65788]: _type = "Task" [ 1197.383976] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.387540] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a16517-633f-b726-34af-d815dec087d6, 'name': SearchDatastore_Task, 'duration_secs': 0.011567} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.391410] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.391600] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1197.391830] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.402286] env[65788]: DEBUG nova.compute.manager [req-43c1cd0c-5971-4c5c-aa38-8630bbe32e06 req-c65e146b-02e5-423d-9a30-3ab72a46ef26 service nova] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Detach interface failed, port_id=731c93dc-1d06-4943-bac0-fc800f9f449d, reason: Instance 50c70e36-a3dc-4e8c-8fe6-74039d1828bf could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1197.406518] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52c90236-bb6a-0a7e-4f77-9ac20620543b, 'name': SearchDatastore_Task, 'duration_secs': 0.011761} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.407447] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42c071b7-c8ec-4a2a-8ffb-14cf027f9829 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.414501] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1197.414501] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff096a-5af4-7041-b3aa-a1e0aeb7d3c0" [ 1197.414501] env[65788]: _type = "Task" [ 1197.414501] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.426446] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff096a-5af4-7041-b3aa-a1e0aeb7d3c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.446181] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663314, 'name': PowerOffVM_Task, 'duration_secs': 0.238655} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.446491] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1197.447381] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6de6bf2-8082-49ef-88ab-bc14d9de3438 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.468654] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3648521e-62f2-4a01-b678-7022973260c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.690864] env[65788]: DEBUG oslo_concurrency.lockutils [None req-cbc13ac4-b582-4bf0-b8c6-4395e10797f5 tempest-ServerActionsTestOtherB-29394606 tempest-ServerActionsTestOtherB-29394606-project-member] Lock "b9d8a32e-1f0f-405c-9543-9e53046d248e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.926s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1197.823424] env[65788]: INFO nova.compute.manager [-] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Took 1.54 seconds to deallocate network for instance. [ 1197.927346] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52ff096a-5af4-7041-b3aa-a1e0aeb7d3c0, 'name': SearchDatastore_Task, 'duration_secs': 0.010342} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.927617] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.927884] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 0636e867-8214-473e-9171-af66a53818a7/0636e867-8214-473e-9171-af66a53818a7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1197.928186] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1197.928373] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1197.928594] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e272488-422e-42f6-8099-650869e3a56a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.930801] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87802b62-21e1-4a95-9c27-7ea6b06650c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.938661] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1197.938661] env[65788]: value = "task-4663315" [ 1197.938661] env[65788]: _type = "Task" [ 1197.938661] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.943231] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1197.943410] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1197.944619] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f34d6d1e-6e00-4906-bc32-f09a61c3516d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.950292] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663315, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.954637] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1197.954637] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525c5052-fe9f-c924-4b05-4dc357e6bd11" [ 1197.954637] env[65788]: _type = "Task" [ 1197.954637] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.964104] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525c5052-fe9f-c924-4b05-4dc357e6bd11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.981216] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Creating Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1197.981583] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-129a8525-d4c4-486f-86d0-75d20e5bdd7b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.990386] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1197.990386] env[65788]: value = "task-4663316" [ 1197.990386] env[65788]: _type = "Task" [ 1197.990386] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.001395] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663316, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.257090] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.257400] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 27ed2344-c42b-46bc-b51a-20821f67cbf0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.257400] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 50c70e36-a3dc-4e8c-8fe6-74039d1828bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.257992] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance bdd09116-31de-491d-b129-c117f898881e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.257992] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f8c57215-ade6-424b-be8c-075998a764af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.257992] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 0636e867-8214-473e-9171-af66a53818a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.257992] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance d15953f8-f74a-422a-867d-fa835bb479ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1198.258661] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1198.258661] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=100GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '7', 'num_vm_active': '5', 'num_task_None': '3', 'num_os_type_None': '7', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '1', 'io_workload': '2', 'num_proj_009f31c7d5bc4d369a8b96e2aa01117a': '2', 'num_task_deleting': '1', 'num_proj_7aa1dec53aa4436daa0f3223b92f9b79': '1', 'num_task_shelving': '1', 'num_proj_94977b75768444138fb914cabddc57c5': '1', 'num_proj_676963c25a3741998b10790689242dbb': '1', 'num_vm_building': '2', 'num_task_spawning': '2', 'num_proj_3326bb2dfdd54781b984431df7ae4609': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1198.331588] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.379513] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f6148a-76ea-4406-806a-854bc73b8d33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.391682] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230ef3e8-a599-4397-9bc2-ef3267ab4075 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.431486] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0d62d3-a5cb-4ef0-a495-3051c3ad4bf4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.447029] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481a3299-3dd8-41e3-bda1-9a6dc5f5e7cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.455874] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663315, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.469679] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1198.479576] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525c5052-fe9f-c924-4b05-4dc357e6bd11, 'name': SearchDatastore_Task, 'duration_secs': 0.011831} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.481501] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2a20c84-95bd-4255-901a-1de09b506486 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.490721] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1198.490721] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a10afe-4c63-2a6b-461b-31ea65a66ba7" [ 1198.490721] env[65788]: _type = "Task" [ 1198.490721] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.506059] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a10afe-4c63-2a6b-461b-31ea65a66ba7, 'name': SearchDatastore_Task, 'duration_secs': 0.011085} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.510024] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1198.510318] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] d15953f8-f74a-422a-867d-fa835bb479ff/d15953f8-f74a-422a-867d-fa835bb479ff.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1198.511017] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663316, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.511272] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04a0e327-c41d-4c47-836a-9be3cfdb1f09 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.522499] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1198.522499] env[65788]: value = "task-4663318" [ 1198.522499] env[65788]: _type = "Task" [ 1198.522499] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.532507] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663318, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.953050] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663315, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534653} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.954129] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 0636e867-8214-473e-9171-af66a53818a7/0636e867-8214-473e-9171-af66a53818a7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1198.954129] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1198.954129] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1fe0416-ecb3-4416-be3c-0a7f9e7563ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.963769] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1198.963769] env[65788]: value = "task-4663319" [ 1198.963769] env[65788]: _type = "Task" [ 1198.963769] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.975171] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1198.982032] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663319, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.007236] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663316, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.036081] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663318, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.474387] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663319, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103095} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.474833] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1199.475525] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a562e8-b19e-4bcd-8991-a5f6dc22235f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.480073] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1199.480266] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.258s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.480519] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.149s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.480742] env[65788]: DEBUG nova.objects.instance [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'resources' on Instance uuid 50c70e36-a3dc-4e8c-8fe6-74039d1828bf {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.501933] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] 0636e867-8214-473e-9171-af66a53818a7/0636e867-8214-473e-9171-af66a53818a7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1199.507200] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91f96c99-ed37-4f20-b2d1-52f759f4ef08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.531321] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663316, 'name': CreateSnapshot_Task, 'duration_secs': 1.24566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.533123] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Created Snapshot of the VM instance {{(pid=65788) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1199.533466] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1199.533466] env[65788]: value = "task-4663320" [ 1199.533466] env[65788]: _type = "Task" [ 1199.533466] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.534574] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43649ec7-5421-4abc-ad6f-b1e28d1f362b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.545319] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663318, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634294} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.545982] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] d15953f8-f74a-422a-867d-fa835bb479ff/d15953f8-f74a-422a-867d-fa835bb479ff.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1199.546254] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1199.546561] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b45f40a-f433-4acd-aa7e-8dcd867b63c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.556797] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663320, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.565205] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1199.565205] env[65788]: value = "task-4663321" [ 1199.565205] env[65788]: _type = "Task" [ 1199.565205] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.575609] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663321, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.052987] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663320, 'name': ReconfigVM_Task, 'duration_secs': 0.318653} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.053376] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Reconfigured VM instance instance-00000074 to attach disk [datastore2] 0636e867-8214-473e-9171-af66a53818a7/0636e867-8214-473e-9171-af66a53818a7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1200.053974] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68cafd60-ef63-456f-b40b-484fe9355f79 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.071371] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Creating linked-clone VM from snapshot {{(pid=65788) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1200.071932] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1200.071932] env[65788]: value = "task-4663322" [ 1200.071932] env[65788]: _type = "Task" [ 1200.071932] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.075740] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-201ece9d-c7a6-4179-a1e1-a406780a6b22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.102159] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663321, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072297} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.102159] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663322, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.102159] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1200.102159] env[65788]: value = "task-4663323" [ 1200.102159] env[65788]: _type = "Task" [ 1200.102159] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.102617] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1200.103873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40ec169-097f-4c87-b588-2ce149a49992 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.119592] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663323, 'name': CloneVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.140158] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] d15953f8-f74a-422a-867d-fa835bb479ff/d15953f8-f74a-422a-867d-fa835bb479ff.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.142469] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed9fdc36-a204-4d8f-af3c-6732f855fb66 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.164585] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1200.164585] env[65788]: value = "task-4663324" [ 1200.164585] env[65788]: _type = "Task" [ 1200.164585] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.181347] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663324, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.200457] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af09deb7-25c7-4b76-b796-0a7691142bb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.211153] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e97847-b0ff-4d30-a66f-a6147bdb1574 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.254948] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00587de6-2105-4472-8535-d678251442b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.268078] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca21e76-e96c-45e5-a92d-80bafe6662a2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.293213] env[65788]: DEBUG nova.compute.provider_tree [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.590397] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663322, 'name': Rename_Task, 'duration_secs': 0.177727} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.590778] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1200.590903] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2c9966b-d15f-41e6-8f94-cb5098fb3508 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.598639] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1200.598639] env[65788]: value = "task-4663325" [ 1200.598639] env[65788]: _type = "Task" [ 1200.598639] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.610493] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663325, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.616383] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663323, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.676018] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663324, 'name': ReconfigVM_Task, 'duration_secs': 0.315401} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.676347] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Reconfigured VM instance instance-00000075 to attach disk [datastore2] d15953f8-f74a-422a-867d-fa835bb479ff/d15953f8-f74a-422a-867d-fa835bb479ff.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1200.677022] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53e52e2a-3189-4091-94a4-fa3f1787e8c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.686244] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1200.686244] env[65788]: value = "task-4663326" [ 1200.686244] env[65788]: _type = "Task" [ 1200.686244] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.695268] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663326, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.797883] env[65788]: DEBUG nova.scheduler.client.report [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1201.114206] env[65788]: DEBUG oslo_vmware.api [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663325, 'name': PowerOnVM_Task, 'duration_secs': 0.505645} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.114936] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1201.115272] env[65788]: INFO nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Took 9.17 seconds to spawn the instance on the hypervisor. [ 1201.115596] env[65788]: DEBUG nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1201.116605] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adc13eb-0c1b-43e9-956a-a9087905ce84 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.123564] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663323, 'name': CloneVM_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.198678] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663326, 'name': Rename_Task, 'duration_secs': 0.198986} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.199022] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1201.199324] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14a02c5a-8b49-4a01-b75c-d83fc86ae623 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.208180] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1201.208180] env[65788]: value = "task-4663327" [ 1201.208180] env[65788]: _type = "Task" [ 1201.208180] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.220066] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663327, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.303935] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1201.331705] env[65788]: INFO nova.scheduler.client.report [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted allocations for instance 50c70e36-a3dc-4e8c-8fe6-74039d1828bf [ 1201.615509] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663323, 'name': CloneVM_Task, 'duration_secs': 1.491229} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.615509] env[65788]: INFO nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Created linked-clone VM from snapshot [ 1201.616212] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c27b828-9d53-4b66-8b26-e88ff228f6fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.624763] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Uploading image e741c886-f6dd-49f9-b52f-b3ee9f0d3301 {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1201.638512] env[65788]: INFO nova.compute.manager [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Took 17.23 seconds to build instance. [ 1201.651464] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1201.651464] env[65788]: value = "vm-910442" [ 1201.651464] env[65788]: _type = "VirtualMachine" [ 1201.651464] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1201.651924] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7d4958e0-c5cd-46f5-b40e-89093a0142e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.659638] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease: (returnval){ [ 1201.659638] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b3243-6b45-4e61-69af-860ee35300a3" [ 1201.659638] env[65788]: _type = "HttpNfcLease" [ 1201.659638] env[65788]: } obtained for exporting VM: (result){ [ 1201.659638] env[65788]: value = "vm-910442" [ 1201.659638] env[65788]: _type = "VirtualMachine" [ 1201.659638] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1201.659934] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the lease: (returnval){ [ 1201.659934] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b3243-6b45-4e61-69af-860ee35300a3" [ 1201.659934] env[65788]: _type = "HttpNfcLease" [ 1201.659934] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1201.668357] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1201.668357] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b3243-6b45-4e61-69af-860ee35300a3" [ 1201.668357] env[65788]: _type = "HttpNfcLease" [ 1201.668357] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1201.719980] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663327, 'name': PowerOnVM_Task} progress is 88%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.839619] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0968132b-73f3-49f3-bdd2-ef012d036f00 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "50c70e36-a3dc-4e8c-8fe6-74039d1828bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.220s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.142673] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12f7c810-2456-43b7-8a55-1e4412db1119 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "0636e867-8214-473e-9171-af66a53818a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.748s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.171787] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1202.171787] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b3243-6b45-4e61-69af-860ee35300a3" [ 1202.171787] env[65788]: _type = "HttpNfcLease" [ 1202.171787] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1202.173444] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1202.173444] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b3243-6b45-4e61-69af-860ee35300a3" [ 1202.173444] env[65788]: _type = "HttpNfcLease" [ 1202.173444] env[65788]: }. {{(pid=65788) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1202.173830] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d344fec-7fae-408c-8a40-3b7cbf1f4ecb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.183949] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e3742f-1d5f-659d-7ac5-f1bbb565ea13/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1202.184155] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e3742f-1d5f-659d-7ac5-f1bbb565ea13/disk-0.vmdk for reading. {{(pid=65788) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1202.254758] env[65788]: DEBUG oslo_vmware.api [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663327, 'name': PowerOnVM_Task, 'duration_secs': 0.82104} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.255175] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1202.255431] env[65788]: INFO nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Took 8.06 seconds to spawn the instance on the hypervisor. [ 1202.255670] env[65788]: DEBUG nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1202.256616] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87343cff-fdd6-4422-824f-03678ee68a47 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.276948] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2d072f3e-63dd-4a4d-9ca1-a0e98d0f6474 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.776630] env[65788]: INFO nova.compute.manager [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Took 14.06 seconds to build instance. [ 1202.863316] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "0636e867-8214-473e-9171-af66a53818a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.863661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "0636e867-8214-473e-9171-af66a53818a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.865122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "0636e867-8214-473e-9171-af66a53818a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.865376] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "0636e867-8214-473e-9171-af66a53818a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.865712] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "0636e867-8214-473e-9171-af66a53818a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.871112] env[65788]: INFO nova.compute.manager [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Terminating instance [ 1203.279192] env[65788]: DEBUG oslo_concurrency.lockutils [None req-793f618d-5cd7-4a7b-a123-71812b641f24 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "d15953f8-f74a-422a-867d-fa835bb479ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.571s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.378023] env[65788]: DEBUG nova.compute.manager [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1203.378023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1203.378023] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b20f85e-1903-4d63-8c16-ae23201d1635 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.391603] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.391603] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f2bead3-a61f-47e5-a11b-33352487acc3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.403215] env[65788]: DEBUG oslo_vmware.api [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1203.403215] env[65788]: value = "task-4663329" [ 1203.403215] env[65788]: _type = "Task" [ 1203.403215] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.426603] env[65788]: DEBUG oslo_vmware.api [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.916177] env[65788]: DEBUG oslo_vmware.api [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663329, 'name': PowerOffVM_Task, 'duration_secs': 0.355978} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.917088] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.918196] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.918196] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67c83d49-6e0c-4f42-8803-4dacb8bbb347 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.004052] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1204.004052] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1204.004052] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Deleting the datastore file [datastore2] 0636e867-8214-473e-9171-af66a53818a7 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1204.004052] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8337f6cd-5c1c-4e1c-be83-9f44509a1214 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.012903] env[65788]: DEBUG oslo_vmware.api [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for the task: (returnval){ [ 1204.012903] env[65788]: value = "task-4663331" [ 1204.012903] env[65788]: _type = "Task" [ 1204.012903] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.025765] env[65788]: DEBUG oslo_vmware.api [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.524971] env[65788]: DEBUG oslo_vmware.api [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Task: {'id': task-4663331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248141} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.525200] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1204.525388] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1204.525597] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1204.525783] env[65788]: INFO nova.compute.manager [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] [instance: 0636e867-8214-473e-9171-af66a53818a7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1204.526097] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1204.526274] env[65788]: DEBUG nova.compute.manager [-] [instance: 0636e867-8214-473e-9171-af66a53818a7] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1204.526371] env[65788]: DEBUG nova.network.neutron [-] [instance: 0636e867-8214-473e-9171-af66a53818a7] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1204.526619] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1204.527178] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1204.527438] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1204.577269] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1204.736141] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "152b3854-f2c6-471b-ac89-b32c51cfb762" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.736404] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.951651] env[65788]: DEBUG nova.compute.manager [req-5de58ec6-8dad-491e-8a45-8f0a4fe29b21 req-456d9505-7072-4c70-9ff9-e8b24985f8be service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Received event network-vif-deleted-b7dbcca3-428b-4b01-9903-f12350d7be4d {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1204.951651] env[65788]: INFO nova.compute.manager [req-5de58ec6-8dad-491e-8a45-8f0a4fe29b21 req-456d9505-7072-4c70-9ff9-e8b24985f8be service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Neutron deleted interface b7dbcca3-428b-4b01-9903-f12350d7be4d; detaching it from the instance and deleting it from the info cache [ 1204.951651] env[65788]: DEBUG nova.network.neutron [req-5de58ec6-8dad-491e-8a45-8f0a4fe29b21 req-456d9505-7072-4c70-9ff9-e8b24985f8be service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1205.071402] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.071778] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.239786] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1205.354697] env[65788]: DEBUG nova.network.neutron [-] [instance: 0636e867-8214-473e-9171-af66a53818a7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1205.453923] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-385088be-4c04-4962-9dd5-8868e30b0c77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.466668] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d568ad6e-adba-44ae-a3d5-fde9937da133 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.503040] env[65788]: DEBUG nova.compute.manager [req-5de58ec6-8dad-491e-8a45-8f0a4fe29b21 req-456d9505-7072-4c70-9ff9-e8b24985f8be service nova] [instance: 0636e867-8214-473e-9171-af66a53818a7] Detach interface failed, port_id=b7dbcca3-428b-4b01-9903-f12350d7be4d, reason: Instance 0636e867-8214-473e-9171-af66a53818a7 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1205.548206] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-f8c57215-ade6-424b-be8c-075998a764af-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.548348] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.548674] env[65788]: DEBUG nova.objects.instance [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'flavor' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.575024] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1205.765249] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.765594] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.767199] env[65788]: INFO nova.compute.claims [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1205.857539] env[65788]: INFO nova.compute.manager [-] [instance: 0636e867-8214-473e-9171-af66a53818a7] Took 1.33 seconds to deallocate network for instance. [ 1206.052450] env[65788]: WARNING neutronclient.v2_0.client [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.053311] env[65788]: WARNING openstack [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.053726] env[65788]: WARNING openstack [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.059117] env[65788]: DEBUG nova.objects.instance [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'pci_requests' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1206.099325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.364791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.561997] env[65788]: DEBUG nova.objects.base [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1206.562366] env[65788]: DEBUG nova.network.neutron [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1206.562692] env[65788]: WARNING neutronclient.v2_0.client [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.563018] env[65788]: WARNING neutronclient.v2_0.client [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1206.563648] env[65788]: WARNING openstack [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1206.564026] env[65788]: WARNING openstack [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1206.651380] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e98f9b9b-69f1-4f98-8498-07e3b7479030 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.103s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.897235] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f336adbe-2788-47da-ad4a-e3c3d6dc5031 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.905935] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde0105c-ca97-4526-a428-64d0843a2c80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.938785] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b736ac-7fa6-4e7b-831a-deb72966e222 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.947218] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a202936a-30c3-4a04-8624-e9471035fcd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.961758] env[65788]: DEBUG nova.compute.provider_tree [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.347461] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "222d1456-baad-42ee-bdef-020fb66ebe41" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.347805] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "222d1456-baad-42ee-bdef-020fb66ebe41" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.465766] env[65788]: DEBUG nova.scheduler.client.report [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1207.849813] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1207.971389] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.972108] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1207.974927] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.876s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.976478] env[65788]: INFO nova.compute.claims [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1208.375404] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1208.481433] env[65788]: DEBUG nova.compute.utils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1208.485163] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1208.485388] env[65788]: DEBUG nova.network.neutron [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1208.485732] env[65788]: WARNING neutronclient.v2_0.client [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.486053] env[65788]: WARNING neutronclient.v2_0.client [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1208.486648] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1208.487058] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1208.537946] env[65788]: DEBUG nova.policy [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1208.590868] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-f8c57215-ade6-424b-be8c-075998a764af-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1208.591169] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1208.591490] env[65788]: DEBUG nova.objects.instance [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'flavor' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.854342] env[65788]: DEBUG nova.network.neutron [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Successfully created port: c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1208.986035] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1209.098775] env[65788]: WARNING neutronclient.v2_0.client [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.099553] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.099909] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.152950] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2381c2-2d91-4f2c-9174-860a649e288b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.162040] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7bae48-665b-4366-90b1-41d193213623 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.200990] env[65788]: DEBUG nova.objects.instance [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'pci_requests' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.204324] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38324e7f-3fb8-4155-8356-35b6e3f050ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.214168] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4554710a-f497-43f6-84ea-0030ed05fedc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.232713] env[65788]: DEBUG nova.compute.provider_tree [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.704376] env[65788]: DEBUG nova.objects.base [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1209.704811] env[65788]: DEBUG nova.network.neutron [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1209.705042] env[65788]: WARNING neutronclient.v2_0.client [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.705383] env[65788]: WARNING neutronclient.v2_0.client [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1209.706096] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1209.706493] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1209.736984] env[65788]: DEBUG nova.scheduler.client.report [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1209.755378] env[65788]: DEBUG nova.policy [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1209.877843] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e3742f-1d5f-659d-7ac5-f1bbb565ea13/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1209.878873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166b76ef-712a-41a8-abb3-0c24f74700ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.886989] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e3742f-1d5f-659d-7ac5-f1bbb565ea13/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1209.887260] env[65788]: ERROR oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e3742f-1d5f-659d-7ac5-f1bbb565ea13/disk-0.vmdk due to incomplete transfer. [ 1209.887458] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2bfd19d8-db90-4910-8f28-0c2972fc8bfa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.897608] env[65788]: DEBUG oslo_vmware.rw_handles [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e3742f-1d5f-659d-7ac5-f1bbb565ea13/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1209.897870] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Uploaded image e741c886-f6dd-49f9-b52f-b3ee9f0d3301 to the Glance image server {{(pid=65788) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1209.900500] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Destroying the VM {{(pid=65788) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1209.900800] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d7549fda-246c-4540-bfde-8bbee1ce59a6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.909303] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1209.909303] env[65788]: value = "task-4663332" [ 1209.909303] env[65788]: _type = "Task" [ 1209.909303] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.920059] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663332, 'name': Destroy_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.996809] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1210.021801] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1210.022085] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1210.023055] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1210.023055] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1210.023055] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1210.023055] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1210.023055] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1210.023265] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1210.023411] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1210.023619] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1210.023730] env[65788]: DEBUG nova.virt.hardware [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1210.024689] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc61fa91-4380-435a-84a8-8b84539fd820 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.036179] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0bbfe58-05d8-4815-b79c-953420c09cf0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.086408] env[65788]: DEBUG nova.network.neutron [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Successfully created port: c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1210.242708] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.243407] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1210.246857] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.882s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.247169] env[65788]: DEBUG nova.objects.instance [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lazy-loading 'resources' on Instance uuid 0636e867-8214-473e-9171-af66a53818a7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1210.380057] env[65788]: DEBUG nova.network.neutron [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Successfully updated port: c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1210.419793] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663332, 'name': Destroy_Task, 'duration_secs': 0.387877} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.420079] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Destroyed the VM [ 1210.420325] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Deleting Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1210.420612] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-90fbdd2e-37f3-452a-9910-0ef8c6a37158 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.428540] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1210.428540] env[65788]: value = "task-4663333" [ 1210.428540] env[65788]: _type = "Task" [ 1210.428540] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.437236] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663333, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.504896] env[65788]: DEBUG nova.compute.manager [req-96d0ba04-077d-4966-a9ae-66a0f4153fc7 req-b11074c9-e005-421f-a2ef-3c5f0597d2ac service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Received event network-vif-plugged-c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1210.505166] env[65788]: DEBUG oslo_concurrency.lockutils [req-96d0ba04-077d-4966-a9ae-66a0f4153fc7 req-b11074c9-e005-421f-a2ef-3c5f0597d2ac service nova] Acquiring lock "152b3854-f2c6-471b-ac89-b32c51cfb762-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.505409] env[65788]: DEBUG oslo_concurrency.lockutils [req-96d0ba04-077d-4966-a9ae-66a0f4153fc7 req-b11074c9-e005-421f-a2ef-3c5f0597d2ac service nova] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.505626] env[65788]: DEBUG oslo_concurrency.lockutils [req-96d0ba04-077d-4966-a9ae-66a0f4153fc7 req-b11074c9-e005-421f-a2ef-3c5f0597d2ac service nova] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.505825] env[65788]: DEBUG nova.compute.manager [req-96d0ba04-077d-4966-a9ae-66a0f4153fc7 req-b11074c9-e005-421f-a2ef-3c5f0597d2ac service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] No waiting events found dispatching network-vif-plugged-c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1210.505993] env[65788]: WARNING nova.compute.manager [req-96d0ba04-077d-4966-a9ae-66a0f4153fc7 req-b11074c9-e005-421f-a2ef-3c5f0597d2ac service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Received unexpected event network-vif-plugged-c703390b-d2a4-4b35-b51c-9b66c1714fb2 for instance with vm_state building and task_state spawning. [ 1210.751306] env[65788]: DEBUG nova.compute.utils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1210.756048] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1210.756410] env[65788]: DEBUG nova.network.neutron [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1210.756704] env[65788]: WARNING neutronclient.v2_0.client [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.757058] env[65788]: WARNING neutronclient.v2_0.client [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1210.758476] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1210.759035] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1210.813997] env[65788]: DEBUG nova.policy [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca13693271604f5790bdc8e19aa208a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7aa1dec53aa4436daa0f3223b92f9b79', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1210.883360] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-152b3854-f2c6-471b-ac89-b32c51cfb762" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.883542] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-152b3854-f2c6-471b-ac89-b32c51cfb762" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1210.883721] env[65788]: DEBUG nova.network.neutron [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1210.886391] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af40010a-e912-4529-8189-420931882115 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.896194] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1347f525-7a3e-4eb3-8903-a01099aa50f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.927863] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0b424a-cb0a-49c0-97f3-3ba1127dfc3c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.942569] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d10faae-79bb-4b82-af5c-caa663edff24 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.947446] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663333, 'name': RemoveSnapshot_Task, 'duration_secs': 0.364685} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.947732] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Deleted Snapshot of the VM instance {{(pid=65788) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1210.948023] env[65788]: DEBUG nova.compute.manager [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1210.949377] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0460cba-26d8-451b-82f7-0aaca5eb5146 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.960963] env[65788]: DEBUG nova.compute.provider_tree [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.124631] env[65788]: DEBUG nova.network.neutron [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Successfully created port: f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1211.257145] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1211.391063] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.391516] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.464375] env[65788]: DEBUG nova.scheduler.client.report [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1211.472960] env[65788]: INFO nova.compute.manager [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Shelve offloading [ 1211.619932] env[65788]: DEBUG nova.network.neutron [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Successfully updated port: c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1211.705167] env[65788]: DEBUG nova.network.neutron [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1211.744511] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.744971] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.822216] env[65788]: DEBUG nova.compute.manager [req-d44b1fd1-80f7-4e0b-b6a0-220c373e10bd req-bd1d0ab8-f84a-45bd-83de-defb7e30943e service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-vif-plugged-c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1211.823158] env[65788]: DEBUG oslo_concurrency.lockutils [req-d44b1fd1-80f7-4e0b-b6a0-220c373e10bd req-bd1d0ab8-f84a-45bd-83de-defb7e30943e service nova] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1211.823505] env[65788]: DEBUG oslo_concurrency.lockutils [req-d44b1fd1-80f7-4e0b-b6a0-220c373e10bd req-bd1d0ab8-f84a-45bd-83de-defb7e30943e service nova] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.823607] env[65788]: DEBUG oslo_concurrency.lockutils [req-d44b1fd1-80f7-4e0b-b6a0-220c373e10bd req-bd1d0ab8-f84a-45bd-83de-defb7e30943e service nova] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.823915] env[65788]: DEBUG nova.compute.manager [req-d44b1fd1-80f7-4e0b-b6a0-220c373e10bd req-bd1d0ab8-f84a-45bd-83de-defb7e30943e service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] No waiting events found dispatching network-vif-plugged-c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1211.824107] env[65788]: WARNING nova.compute.manager [req-d44b1fd1-80f7-4e0b-b6a0-220c373e10bd req-bd1d0ab8-f84a-45bd-83de-defb7e30943e service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received unexpected event network-vif-plugged-c720474e-625c-4e3a-8654-dcfc4227f500 for instance with vm_state active and task_state None. [ 1211.831075] env[65788]: WARNING neutronclient.v2_0.client [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1211.831722] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1211.832136] env[65788]: WARNING openstack [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1211.940750] env[65788]: DEBUG nova.network.neutron [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Updating instance_info_cache with network_info: [{"id": "c703390b-d2a4-4b35-b51c-9b66c1714fb2", "address": "fa:16:3e:a5:89:60", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc703390b-d2", "ovs_interfaceid": "c703390b-d2a4-4b35-b51c-9b66c1714fb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1211.969408] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.972266] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.597s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1211.973581] env[65788]: INFO nova.compute.claims [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1211.976963] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1211.976963] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0fac721-31f0-4433-aab9-50d0336d47dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.984882] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1211.984882] env[65788]: value = "task-4663334" [ 1211.984882] env[65788]: _type = "Task" [ 1211.984882] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.995364] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] VM already powered off {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1211.995424] env[65788]: DEBUG nova.compute.manager [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1211.996639] env[65788]: INFO nova.scheduler.client.report [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Deleted allocations for instance 0636e867-8214-473e-9171-af66a53818a7 [ 1211.998526] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d5d44d-0098-4128-9547-14a345afdd6f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.008529] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.008706] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.008876] env[65788]: DEBUG nova.network.neutron [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1212.123208] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.123438] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.123718] env[65788]: DEBUG nova.network.neutron [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1212.268310] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1212.299498] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1212.299772] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1212.299932] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1212.300135] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1212.300273] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1212.300418] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1212.300628] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1212.300789] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1212.300959] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1212.301140] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1212.301318] env[65788]: DEBUG nova.virt.hardware [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1212.302294] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88028534-6da1-4b14-81d8-030dff536528 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.312339] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed151a5a-11db-41a1-9fb3-98eaf464a84d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.443679] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-152b3854-f2c6-471b-ac89-b32c51cfb762" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.444110] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Instance network_info: |[{"id": "c703390b-d2a4-4b35-b51c-9b66c1714fb2", "address": "fa:16:3e:a5:89:60", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc703390b-d2", "ovs_interfaceid": "c703390b-d2a4-4b35-b51c-9b66c1714fb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1212.444672] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:89:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c703390b-d2a4-4b35-b51c-9b66c1714fb2', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1212.452381] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1212.452614] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1212.452870] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ddd0191-747b-4a14-85d2-dd1b511b446d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.477573] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1212.477573] env[65788]: value = "task-4663335" [ 1212.477573] env[65788]: _type = "Task" [ 1212.477573] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.489462] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663335, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.507783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-95ae989d-1470-433b-809f-41d36dff6ae1 tempest-ServerPasswordTestJSON-1934956593 tempest-ServerPasswordTestJSON-1934956593-project-member] Lock "0636e867-8214-473e-9171-af66a53818a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.644s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1212.511989] env[65788]: WARNING neutronclient.v2_0.client [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.513275] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.513476] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.543868] env[65788]: DEBUG nova.compute.manager [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Received event network-changed-c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1212.544173] env[65788]: DEBUG nova.compute.manager [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Refreshing instance network info cache due to event network-changed-c703390b-d2a4-4b35-b51c-9b66c1714fb2. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1212.544365] env[65788]: DEBUG oslo_concurrency.lockutils [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Acquiring lock "refresh_cache-152b3854-f2c6-471b-ac89-b32c51cfb762" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.544459] env[65788]: DEBUG oslo_concurrency.lockutils [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Acquired lock "refresh_cache-152b3854-f2c6-471b-ac89-b32c51cfb762" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.544732] env[65788]: DEBUG nova.network.neutron [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Refreshing network info cache for port c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1212.626918] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.627376] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.648205] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.649076] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.682460] env[65788]: WARNING nova.network.neutron [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1212.701400] env[65788]: DEBUG nova.network.neutron [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Successfully updated port: f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1212.741509] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.742138] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.753732] env[65788]: WARNING neutronclient.v2_0.client [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.754721] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.755143] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.826725] env[65788]: WARNING neutronclient.v2_0.client [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.827931] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.828534] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.867323] env[65788]: DEBUG nova.network.neutron [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1212.952959] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1212.952959] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1212.997582] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663335, 'name': CreateVM_Task, 'duration_secs': 0.423898} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.997871] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1212.998316] env[65788]: WARNING neutronclient.v2_0.client [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1212.998681] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.998837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1212.999247] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1212.999562] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed83d69d-0b2d-41e0-8335-976df76da1ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.005749] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1213.005749] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d51d3c-3305-bb1b-10a8-a4c3064ab17e" [ 1213.005749] env[65788]: _type = "Task" [ 1213.005749] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.019625] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d51d3c-3305-bb1b-10a8-a4c3064ab17e, 'name': SearchDatastore_Task, 'duration_secs': 0.011105} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.020972] env[65788]: WARNING neutronclient.v2_0.client [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.021626] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.021985] env[65788]: WARNING openstack [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.030131] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.030410] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1213.030656] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.030798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.030991] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1213.034256] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b28d8a2f-3470-4d4c-b5cf-6d9ec1caba9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.045164] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1213.045399] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1213.048872] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b495c45-6b1a-4206-bd4c-111e7886b4d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.051914] env[65788]: WARNING neutronclient.v2_0.client [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.052594] env[65788]: WARNING openstack [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.052953] env[65788]: WARNING openstack [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.064295] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1213.064295] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f1e75b-8db2-cca4-9f07-d5e64434e77f" [ 1213.064295] env[65788]: _type = "Task" [ 1213.064295] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.081363] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f1e75b-8db2-cca4-9f07-d5e64434e77f, 'name': SearchDatastore_Task, 'duration_secs': 0.010667} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.082220] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c92ac6d-dc76-4dce-aaf3-b6d81f1de280 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.092485] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1213.092485] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e7ca94-ca71-30c8-dc72-e6b4cb9d109f" [ 1213.092485] env[65788]: _type = "Task" [ 1213.092485] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.110042] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52e7ca94-ca71-30c8-dc72-e6b4cb9d109f, 'name': SearchDatastore_Task, 'duration_secs': 0.011592} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.110336] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.110599] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 152b3854-f2c6-471b-ac89-b32c51cfb762/152b3854-f2c6-471b-ac89-b32c51cfb762.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1213.110887] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8f07974-efee-40c9-9e57-696878d818a4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.120727] env[65788]: DEBUG nova.network.neutron [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c720474e-625c-4e3a-8654-dcfc4227f500", "address": "fa:16:3e:23:6d:d3", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc720474e-62", "ovs_interfaceid": "c720474e-625c-4e3a-8654-dcfc4227f500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.123458] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1213.123458] env[65788]: value = "task-4663336" [ 1213.123458] env[65788]: _type = "Task" [ 1213.123458] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.137277] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663336, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.146035] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497a007f-616b-418b-abc4-913f7bc1cbad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.155553] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff24398-629f-4402-9de8-47027f431b31 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.162303] env[65788]: WARNING openstack [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.162726] env[65788]: WARNING openstack [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.200530] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9acb3aa-d219-4ec4-87e4-f932da65d40c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.209176] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.209395] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.209606] env[65788]: DEBUG nova.network.neutron [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1213.212060] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f50f935-a6bd-4123-a6a4-27ec3059dd33 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.233578] env[65788]: DEBUG nova.compute.provider_tree [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1213.300311] env[65788]: WARNING neutronclient.v2_0.client [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.301636] env[65788]: WARNING openstack [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.302358] env[65788]: WARNING openstack [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.371858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.372348] env[65788]: WARNING neutronclient.v2_0.client [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.373243] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.373723] env[65788]: WARNING openstack [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.379923] env[65788]: WARNING neutronclient.v2_0.client [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1213.398219] env[65788]: DEBUG nova.network.neutron [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Updated VIF entry in instance network info cache for port c703390b-d2a4-4b35-b51c-9b66c1714fb2. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1213.398844] env[65788]: DEBUG nova.network.neutron [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Updating instance_info_cache with network_info: [{"id": "c703390b-d2a4-4b35-b51c-9b66c1714fb2", "address": "fa:16:3e:a5:89:60", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc703390b-d2", "ovs_interfaceid": "c703390b-d2a4-4b35-b51c-9b66c1714fb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1213.625232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.626348] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.626578] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.628120] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6704fe16-41bb-426e-811a-57d4c84afbec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.647448] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663336, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.668359] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1213.668831] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1213.669173] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1213.669524] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1213.669814] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1213.670118] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1213.670474] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1213.670724] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1213.673521] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1213.673521] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1213.673521] env[65788]: DEBUG nova.virt.hardware [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1213.683409] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfiguring VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1213.684654] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c369575-24b8-4b4e-a487-ace95941ec82 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.716818] env[65788]: DEBUG oslo_vmware.api [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1213.716818] env[65788]: value = "task-4663337" [ 1213.716818] env[65788]: _type = "Task" [ 1213.716818] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.724279] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.724776] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.742790] env[65788]: DEBUG nova.scheduler.client.report [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1213.756403] env[65788]: DEBUG oslo_vmware.api [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663337, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.859067] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1213.860044] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b82050d-2f5d-4fc0-89f4-669ccf71b2e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.870528] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1213.870925] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b1b1765-a348-4044-8573-47378267f660 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.875501] env[65788]: DEBUG nova.compute.manager [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-changed-c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1213.875786] env[65788]: DEBUG nova.compute.manager [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing instance network info cache due to event network-changed-c720474e-625c-4e3a-8654-dcfc4227f500. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1213.876064] env[65788]: DEBUG oslo_concurrency.lockutils [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.876250] env[65788]: DEBUG oslo_concurrency.lockutils [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1213.876448] env[65788]: DEBUG nova.network.neutron [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing network info cache for port c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1213.892794] env[65788]: DEBUG nova.network.neutron [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1213.901560] env[65788]: DEBUG oslo_concurrency.lockutils [req-4b05b4c7-4e0a-4ecf-a519-675e57e16453 req-6e0dd779-54aa-45a7-bcbc-549a990be819 service nova] Releasing lock "refresh_cache-152b3854-f2c6-471b-ac89-b32c51cfb762" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1213.913722] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1213.914153] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1213.951152] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1213.951349] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1213.951538] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleting the datastore file [datastore2] bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1213.951840] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-463b2da9-0d8f-4a0c-860b-72693568033d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.966388] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1213.966388] env[65788]: value = "task-4663339" [ 1213.966388] env[65788]: _type = "Task" [ 1213.966388] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.978556] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.010112] env[65788]: WARNING neutronclient.v2_0.client [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.010810] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.011174] env[65788]: WARNING openstack [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.137229] env[65788]: DEBUG nova.network.neutron [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updating instance_info_cache with network_info: [{"id": "f42a00fc-23eb-4df0-a834-72750f41e45f", "address": "fa:16:3e:59:ac:ac", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42a00fc-23", "ovs_interfaceid": "f42a00fc-23eb-4df0-a834-72750f41e45f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1214.148767] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663336, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621608} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.148767] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 152b3854-f2c6-471b-ac89-b32c51cfb762/152b3854-f2c6-471b-ac89-b32c51cfb762.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1214.148767] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1214.148767] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86423474-3484-4c86-9d88-6e0ce7d09dae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.157493] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1214.157493] env[65788]: value = "task-4663340" [ 1214.157493] env[65788]: _type = "Task" [ 1214.157493] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.169585] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663340, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.230838] env[65788]: DEBUG oslo_vmware.api [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663337, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.249957] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.250553] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1214.380067] env[65788]: WARNING neutronclient.v2_0.client [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.381298] env[65788]: WARNING openstack [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.381703] env[65788]: WARNING openstack [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.479617] env[65788]: DEBUG oslo_vmware.api [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.406185} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.479813] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1214.479952] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1214.480195] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1214.507446] env[65788]: INFO nova.scheduler.client.report [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted allocations for instance bdd09116-31de-491d-b129-c117f898881e [ 1214.524591] env[65788]: WARNING openstack [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.525107] env[65788]: WARNING openstack [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.594612] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Received event network-vif-plugged-f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1214.594914] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1214.595131] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1214.595290] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.595450] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] No waiting events found dispatching network-vif-plugged-f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1214.595622] env[65788]: WARNING nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Received unexpected event network-vif-plugged-f42a00fc-23eb-4df0-a834-72750f41e45f for instance with vm_state building and task_state spawning. [ 1214.595800] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Received event network-changed-f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1214.595959] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Refreshing instance network info cache due to event network-changed-f42a00fc-23eb-4df0-a834-72750f41e45f. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1214.596120] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Acquiring lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.607791] env[65788]: WARNING neutronclient.v2_0.client [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.608536] env[65788]: WARNING openstack [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.608937] env[65788]: WARNING openstack [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.640711] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1214.642219] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Instance network_info: |[{"id": "f42a00fc-23eb-4df0-a834-72750f41e45f", "address": "fa:16:3e:59:ac:ac", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42a00fc-23", "ovs_interfaceid": "f42a00fc-23eb-4df0-a834-72750f41e45f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1214.642219] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Acquired lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.642219] env[65788]: DEBUG nova.network.neutron [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Refreshing network info cache for port f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1214.643218] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:ac:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '35fcdc55-dc29-451b-ad56-3a03b044dc81', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f42a00fc-23eb-4df0-a834-72750f41e45f', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1214.651927] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1214.653110] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1214.657643] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f654f89-1341-43fe-a28d-3efc9eec2250 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.685366] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080502} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.686984] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1214.687307] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1214.687307] env[65788]: value = "task-4663341" [ 1214.687307] env[65788]: _type = "Task" [ 1214.687307] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.688047] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93aff90e-e1fa-4106-bc3c-5cb59559c281 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.700342] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663341, 'name': CreateVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.724027] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 152b3854-f2c6-471b-ac89-b32c51cfb762/152b3854-f2c6-471b-ac89-b32c51cfb762.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1214.726951] env[65788]: DEBUG nova.network.neutron [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updated VIF entry in instance network info cache for port c720474e-625c-4e3a-8654-dcfc4227f500. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1214.727566] env[65788]: DEBUG nova.network.neutron [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c720474e-625c-4e3a-8654-dcfc4227f500", "address": "fa:16:3e:23:6d:d3", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc720474e-62", "ovs_interfaceid": "c720474e-625c-4e3a-8654-dcfc4227f500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1214.729155] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5163a666-4800-48f4-8afc-4221c9c9b25a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.759382] env[65788]: DEBUG nova.compute.utils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1214.762984] env[65788]: DEBUG oslo_vmware.api [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663337, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.763975] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1214.764321] env[65788]: DEBUG nova.network.neutron [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1214.764860] env[65788]: WARNING neutronclient.v2_0.client [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.765350] env[65788]: WARNING neutronclient.v2_0.client [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1214.766057] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1214.769022] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1214.777157] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1214.777157] env[65788]: value = "task-4663342" [ 1214.777157] env[65788]: _type = "Task" [ 1214.777157] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.792895] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663342, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.834606] env[65788]: DEBUG nova.policy [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02eddcb388a431abdff72ad3995dda1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69c8fb9f477e45068cd7c1e4defb64a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1215.011936] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.012443] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.012607] env[65788]: DEBUG nova.objects.instance [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'resources' on Instance uuid bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.154242] env[65788]: WARNING neutronclient.v2_0.client [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.154649] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.154950] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.201738] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663341, 'name': CreateVM_Task, 'duration_secs': 0.384763} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.201874] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1215.202419] env[65788]: WARNING neutronclient.v2_0.client [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.202819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.202982] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1215.203372] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1215.206158] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-756cb7af-a471-4683-abff-e8d55ea59acb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.212997] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1215.212997] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b6b48-7edc-52f8-6b6e-96efaf89c14b" [ 1215.212997] env[65788]: _type = "Task" [ 1215.212997] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.225263] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b6b48-7edc-52f8-6b6e-96efaf89c14b, 'name': SearchDatastore_Task, 'duration_secs': 0.010068} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.225636] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.225879] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1215.226132] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.226283] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1215.226464] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1215.226795] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c482d786-9bbf-4672-9b1c-e858339605bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.237561] env[65788]: DEBUG nova.network.neutron [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Successfully created port: fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1215.244618] env[65788]: DEBUG oslo_vmware.api [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663337, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.244618] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1215.244837] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1215.245926] env[65788]: DEBUG oslo_concurrency.lockutils [req-687b1408-85bb-4969-a629-ee7b1ea31593 req-9fdd22bf-38a7-4d8c-8240-eb7f1aa3bb1b service nova] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.246450] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c887ae1-9348-42b5-ab8a-b6a1e44297b6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.252668] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1215.252668] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cf3d48-15d7-c1b8-4d89-25c99b3ae372" [ 1215.252668] env[65788]: _type = "Task" [ 1215.252668] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.262336] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cf3d48-15d7-c1b8-4d89-25c99b3ae372, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.267199] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1215.287286] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663342, 'name': ReconfigVM_Task, 'duration_secs': 0.299004} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.287538] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 152b3854-f2c6-471b-ac89-b32c51cfb762/152b3854-f2c6-471b-ac89-b32c51cfb762.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.288252] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c5820b0-cccb-4cf5-ad73-11651aa6f9c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.296020] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1215.296020] env[65788]: value = "task-4663343" [ 1215.296020] env[65788]: _type = "Task" [ 1215.296020] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.297480] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.297945] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.316764] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663343, 'name': Rename_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.398305] env[65788]: WARNING neutronclient.v2_0.client [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.399038] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1215.399382] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1215.515940] env[65788]: DEBUG nova.objects.instance [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'numa_topology' on Instance uuid bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.520854] env[65788]: DEBUG nova.network.neutron [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updated VIF entry in instance network info cache for port f42a00fc-23eb-4df0-a834-72750f41e45f. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1215.521213] env[65788]: DEBUG nova.network.neutron [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updating instance_info_cache with network_info: [{"id": "f42a00fc-23eb-4df0-a834-72750f41e45f", "address": "fa:16:3e:59:ac:ac", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42a00fc-23", "ovs_interfaceid": "f42a00fc-23eb-4df0-a834-72750f41e45f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1215.740217] env[65788]: DEBUG oslo_vmware.api [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663337, 'name': ReconfigVM_Task, 'duration_secs': 1.909568} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.740679] env[65788]: WARNING neutronclient.v2_0.client [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1215.741016] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.741247] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfigured VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1215.765056] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cf3d48-15d7-c1b8-4d89-25c99b3ae372, 'name': SearchDatastore_Task, 'duration_secs': 0.010053} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.765999] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef6e2022-506b-4797-b1be-e6c5e1874e3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.776809] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1215.776809] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d0b203-7fb0-a2f2-8fec-767c1855d763" [ 1215.776809] env[65788]: _type = "Task" [ 1215.776809] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.787761] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d0b203-7fb0-a2f2-8fec-767c1855d763, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.816766] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663343, 'name': Rename_Task, 'duration_secs': 0.320873} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.817069] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1215.817337] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07813507-0c26-49a7-84c1-bcf7697cfafd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.824132] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1215.824132] env[65788]: value = "task-4663344" [ 1215.824132] env[65788]: _type = "Task" [ 1215.824132] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.833210] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663344, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.019369] env[65788]: DEBUG nova.objects.base [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1216.023785] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Releasing lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1216.024051] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-vif-unplugged-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1216.024255] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.024473] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Lock "bdd09116-31de-491d-b129-c117f898881e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.024664] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Lock "bdd09116-31de-491d-b129-c117f898881e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.024842] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] No waiting events found dispatching network-vif-unplugged-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1216.025027] env[65788]: WARNING nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received unexpected event network-vif-unplugged-c6074f79-c6f2-4186-b2c3-809f045737ed for instance with vm_state shelved_offloaded and task_state None. [ 1216.025439] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1216.025439] env[65788]: DEBUG nova.compute.manager [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing instance network info cache due to event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1216.025620] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.025658] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1216.025785] env[65788]: DEBUG nova.network.neutron [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1216.132818] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853c7d13-6508-4a60-84ea-61895665dbf6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.141378] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f18c0c-82b3-418a-9ad5-86f58e8f2fe9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.178626] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b30c9dc-2ed8-42f7-8355-2c033ba4032f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.188531] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655566e7-0937-40ff-bda8-66343cc4df9a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.205145] env[65788]: DEBUG nova.compute.provider_tree [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.246424] env[65788]: DEBUG oslo_concurrency.lockutils [None req-daa5d515-78f8-40cb-8c26-26a33b4d5856 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.655s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1216.282266] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1216.291943] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d0b203-7fb0-a2f2-8fec-767c1855d763, 'name': SearchDatastore_Task, 'duration_secs': 0.041383} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.292239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1216.292589] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0/ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1216.292888] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-83b11ad3-abe2-40ef-998d-878511805e60 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.302634] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1216.302634] env[65788]: value = "task-4663345" [ 1216.302634] env[65788]: _type = "Task" [ 1216.302634] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.314989] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.317641] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1216.317974] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1216.318140] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1216.318314] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1216.318523] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1216.318716] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1216.318954] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1216.319159] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1216.319457] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1216.319557] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1216.319773] env[65788]: DEBUG nova.virt.hardware [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1216.320754] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b0affc-007d-4665-bd22-5087d00c29d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.331733] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e104ca-b0ee-46c8-96cf-cf242e12077b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.339479] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663344, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.528751] env[65788]: WARNING neutronclient.v2_0.client [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1216.529770] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.530303] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.712358] env[65788]: DEBUG nova.scheduler.client.report [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1216.764448] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.765021] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.815853] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663345, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.839074] env[65788]: DEBUG oslo_vmware.api [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663344, 'name': PowerOnVM_Task, 'duration_secs': 0.89149} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.839383] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.839593] env[65788]: INFO nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Took 6.84 seconds to spawn the instance on the hypervisor. [ 1216.840700] env[65788]: DEBUG nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1216.845300] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd3b291-1a2c-4b49-b9a5-8b5789731914 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.859471] env[65788]: DEBUG nova.network.neutron [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Successfully updated port: fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1216.892157] env[65788]: WARNING neutronclient.v2_0.client [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1216.892694] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1216.893071] env[65788]: WARNING openstack [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1216.975430] env[65788]: DEBUG nova.network.neutron [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updated VIF entry in instance network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1216.976023] env[65788]: DEBUG nova.network.neutron [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc6074f79-c6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1217.082332] env[65788]: DEBUG nova.compute.manager [req-2a2963ce-86a0-4b77-b0d6-9accb947f5e1 req-f67b5d0d-fc83-4993-a7be-adcb1379b2a4 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Received event network-vif-plugged-fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1217.083027] env[65788]: DEBUG oslo_concurrency.lockutils [req-2a2963ce-86a0-4b77-b0d6-9accb947f5e1 req-f67b5d0d-fc83-4993-a7be-adcb1379b2a4 service nova] Acquiring lock "222d1456-baad-42ee-bdef-020fb66ebe41-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1217.083325] env[65788]: DEBUG oslo_concurrency.lockutils [req-2a2963ce-86a0-4b77-b0d6-9accb947f5e1 req-f67b5d0d-fc83-4993-a7be-adcb1379b2a4 service nova] Lock "222d1456-baad-42ee-bdef-020fb66ebe41-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1217.083415] env[65788]: DEBUG oslo_concurrency.lockutils [req-2a2963ce-86a0-4b77-b0d6-9accb947f5e1 req-f67b5d0d-fc83-4993-a7be-adcb1379b2a4 service nova] Lock "222d1456-baad-42ee-bdef-020fb66ebe41-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.083709] env[65788]: DEBUG nova.compute.manager [req-2a2963ce-86a0-4b77-b0d6-9accb947f5e1 req-f67b5d0d-fc83-4993-a7be-adcb1379b2a4 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] No waiting events found dispatching network-vif-plugged-fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1217.083854] env[65788]: WARNING nova.compute.manager [req-2a2963ce-86a0-4b77-b0d6-9accb947f5e1 req-f67b5d0d-fc83-4993-a7be-adcb1379b2a4 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Received unexpected event network-vif-plugged-fff8c844-a1fe-4657-9ab0-99fa3a35d346 for instance with vm_state building and task_state spawning. [ 1217.218318] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.206s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.314987] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557667} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.315232] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0/ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1217.315547] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1217.316055] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42433de9-292c-4d31-aa14-1cd01dc4d5ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.324469] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1217.324469] env[65788]: value = "task-4663346" [ 1217.324469] env[65788]: _type = "Task" [ 1217.324469] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.334761] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663346, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.367209] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.367209] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1217.367209] env[65788]: DEBUG nova.network.neutron [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1217.371984] env[65788]: INFO nova.compute.manager [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Took 11.63 seconds to build instance. [ 1217.480078] env[65788]: DEBUG oslo_concurrency.lockutils [req-3a1b7f8e-a5df-47c3-9d19-cbc1b2e9f199 req-d57d4003-ada4-401a-ba6a-b76fa1d2b22a service nova] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1217.728448] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2759b9d5-7dd7-4158-aeee-009773b18a55 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.821s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.836669] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663346, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.169483} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.838225] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1217.841680] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288583cb-975e-4341-9cde-0e1fc2714a84 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.870610] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0/ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1217.872549] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1217.872840] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1217.883673] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75385aca-7a1d-45ea-8f3a-cf4a816c75f9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.911280] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d043ad5b-3c58-4de8-ac46-ec848230abb3 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.175s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.919603] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1217.919603] env[65788]: value = "task-4663347" [ 1217.919603] env[65788]: _type = "Task" [ 1217.919603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.934263] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663347, 'name': ReconfigVM_Task} progress is 10%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.941990] env[65788]: DEBUG nova.network.neutron [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1217.964810] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1217.965528] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.036400] env[65788]: WARNING neutronclient.v2_0.client [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.037171] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.037550] env[65788]: WARNING openstack [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.046463] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "152b3854-f2c6-471b-ac89-b32c51cfb762" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.046703] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.046920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "152b3854-f2c6-471b-ac89-b32c51cfb762-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.047147] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.047310] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.049475] env[65788]: INFO nova.compute.manager [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Terminating instance [ 1218.134553] env[65788]: DEBUG nova.network.neutron [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updating instance_info_cache with network_info: [{"id": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "address": "fa:16:3e:f7:50:eb", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff8c844-a1", "ovs_interfaceid": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1218.261145] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.261423] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.261689] env[65788]: INFO nova.compute.manager [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Unshelving [ 1218.346205] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-f8c57215-ade6-424b-be8c-075998a764af-114649fa-5d38-4198-8cf6-45044b37f249" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1218.346420] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-114649fa-5d38-4198-8cf6-45044b37f249" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1218.346774] env[65788]: DEBUG nova.objects.instance [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'flavor' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.429772] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663347, 'name': ReconfigVM_Task, 'duration_secs': 0.305499} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.430097] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Reconfigured VM instance instance-00000077 to attach disk [datastore2] ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0/ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1218.430755] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5570aed-0087-464f-b517-13cbd035305d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.438332] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1218.438332] env[65788]: value = "task-4663348" [ 1218.438332] env[65788]: _type = "Task" [ 1218.438332] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.447309] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663348, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.554172] env[65788]: DEBUG nova.compute.manager [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1218.554536] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1218.555879] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29c6ded-489d-49f0-a346-90e8beacd5ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.564216] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1218.564488] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88cc0c10-25a0-4cbf-822a-32811f3926e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.571058] env[65788]: DEBUG oslo_vmware.api [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1218.571058] env[65788]: value = "task-4663349" [ 1218.571058] env[65788]: _type = "Task" [ 1218.571058] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.581254] env[65788]: DEBUG oslo_vmware.api [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663349, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.637871] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1218.638275] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance network_info: |[{"id": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "address": "fa:16:3e:f7:50:eb", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff8c844-a1", "ovs_interfaceid": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1218.638775] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:50:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fff8c844-a1fe-4657-9ab0-99fa3a35d346', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1218.646616] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1218.646887] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1218.647184] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b2778ca-c46a-4a54-ac9c-2d02c7ea8a6f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.668295] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1218.668295] env[65788]: value = "task-4663350" [ 1218.668295] env[65788]: _type = "Task" [ 1218.668295] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.676744] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663350, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.857086] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1218.857086] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1218.857086] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1218.954691] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663348, 'name': Rename_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.998555] env[65788]: DEBUG nova.objects.instance [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'pci_requests' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.082748] env[65788]: DEBUG oslo_vmware.api [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663349, 'name': PowerOffVM_Task, 'duration_secs': 0.256664} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.083085] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1219.083266] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1219.083536] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c258b01f-928c-4949-9c12-5a618991a1bb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.168644] env[65788]: DEBUG nova.compute.manager [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Received event network-changed-fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1219.169163] env[65788]: DEBUG nova.compute.manager [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Refreshing instance network info cache due to event network-changed-fff8c844-a1fe-4657-9ab0-99fa3a35d346. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1219.169230] env[65788]: DEBUG oslo_concurrency.lockutils [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Acquiring lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.169835] env[65788]: DEBUG oslo_concurrency.lockutils [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Acquired lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1219.169835] env[65788]: DEBUG nova.network.neutron [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Refreshing network info cache for port fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1219.174405] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1219.174405] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1219.174405] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore2] 152b3854-f2c6-471b-ac89-b32c51cfb762 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1219.178414] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecd84306-052b-4c0e-a097-617c06592034 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.189579] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663350, 'name': CreateVM_Task, 'duration_secs': 0.428837} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.190688] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1219.191086] env[65788]: DEBUG oslo_vmware.api [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1219.191086] env[65788]: value = "task-4663352" [ 1219.191086] env[65788]: _type = "Task" [ 1219.191086] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.191589] env[65788]: WARNING neutronclient.v2_0.client [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.192256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.192256] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1219.192477] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1219.193166] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad8344f0-d320-402f-9f09-505a2b3c07b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.204578] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1219.204578] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fce206-8379-cc74-8ffa-33d11213adb0" [ 1219.204578] env[65788]: _type = "Task" [ 1219.204578] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.207829] env[65788]: DEBUG oslo_vmware.api [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663352, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.220062] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52fce206-8379-cc74-8ffa-33d11213adb0, 'name': SearchDatastore_Task, 'duration_secs': 0.012434} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.220380] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1219.220621] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1219.220863] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.221015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1219.221302] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1219.221599] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06b76596-0481-43ff-b8d2-56e6a80f5fb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.232449] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1219.232686] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1219.233496] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b953520-6dfa-4305-9031-63736c67031b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.240916] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1219.240916] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526dead4-c91a-b12a-f5ee-85bb34eea167" [ 1219.240916] env[65788]: _type = "Task" [ 1219.240916] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.250754] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526dead4-c91a-b12a-f5ee-85bb34eea167, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.290012] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.290326] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.290544] env[65788]: DEBUG nova.objects.instance [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'pci_requests' on Instance uuid bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.450430] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663348, 'name': Rename_Task, 'duration_secs': 0.881525} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.450690] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1219.450970] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eabddc4c-ab3e-4cc4-880a-48f5312737fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.458538] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1219.458538] env[65788]: value = "task-4663353" [ 1219.458538] env[65788]: _type = "Task" [ 1219.458538] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.467324] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663353, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.501447] env[65788]: DEBUG nova.objects.base [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1219.501722] env[65788]: DEBUG nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1219.502122] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.502471] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.503151] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.503529] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.579077] env[65788]: DEBUG nova.policy [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1219.675846] env[65788]: WARNING neutronclient.v2_0.client [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.676678] env[65788]: WARNING openstack [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.677099] env[65788]: WARNING openstack [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.703464] env[65788]: DEBUG oslo_vmware.api [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.359471} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.703699] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1219.703851] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1219.704050] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1219.704230] env[65788]: INFO nova.compute.manager [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1219.704490] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1219.704726] env[65788]: DEBUG nova.compute.manager [-] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1219.704867] env[65788]: DEBUG nova.network.neutron [-] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1219.705194] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.705877] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1219.706219] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1219.746849] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1219.756550] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]526dead4-c91a-b12a-f5ee-85bb34eea167, 'name': SearchDatastore_Task, 'duration_secs': 0.012335} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.757785] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6478092-ca32-4445-be29-75dce2f32aab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.765492] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1219.765492] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d6b919-232a-3c10-5284-10a76c2d507a" [ 1219.765492] env[65788]: _type = "Task" [ 1219.765492] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.778578] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d6b919-232a-3c10-5284-10a76c2d507a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.794894] env[65788]: DEBUG nova.objects.instance [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'numa_topology' on Instance uuid bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.972160] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663353, 'name': PowerOnVM_Task} progress is 89%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.007532] env[65788]: WARNING openstack [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.008134] env[65788]: WARNING openstack [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.058680] env[65788]: DEBUG nova.compute.manager [req-b6848355-f091-4f64-98ad-75d213470bcf req-bcc2c13f-62b0-4998-9164-6d0a7d85a8ce service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Received event network-vif-deleted-c703390b-d2a4-4b35-b51c-9b66c1714fb2 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1220.058889] env[65788]: INFO nova.compute.manager [req-b6848355-f091-4f64-98ad-75d213470bcf req-bcc2c13f-62b0-4998-9164-6d0a7d85a8ce service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Neutron deleted interface c703390b-d2a4-4b35-b51c-9b66c1714fb2; detaching it from the instance and deleting it from the info cache [ 1220.059073] env[65788]: DEBUG nova.network.neutron [req-b6848355-f091-4f64-98ad-75d213470bcf req-bcc2c13f-62b0-4998-9164-6d0a7d85a8ce service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1220.085505] env[65788]: WARNING neutronclient.v2_0.client [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1220.086395] env[65788]: WARNING openstack [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1220.086813] env[65788]: WARNING openstack [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1220.218054] env[65788]: DEBUG nova.network.neutron [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updated VIF entry in instance network info cache for port fff8c844-a1fe-4657-9ab0-99fa3a35d346. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1220.218440] env[65788]: DEBUG nova.network.neutron [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updating instance_info_cache with network_info: [{"id": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "address": "fa:16:3e:f7:50:eb", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff8c844-a1", "ovs_interfaceid": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1220.277029] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d6b919-232a-3c10-5284-10a76c2d507a, 'name': SearchDatastore_Task, 'duration_secs': 0.013446} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.277325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1220.277588] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1220.277853] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de4684bf-d194-455c-a727-de0d23609294 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.285998] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1220.285998] env[65788]: value = "task-4663354" [ 1220.285998] env[65788]: _type = "Task" [ 1220.285998] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.295320] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.297750] env[65788]: INFO nova.compute.claims [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1220.472461] env[65788]: DEBUG oslo_vmware.api [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663353, 'name': PowerOnVM_Task, 'duration_secs': 0.59473} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.472461] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1220.472693] env[65788]: INFO nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1220.472928] env[65788]: DEBUG nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1220.473951] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64da7195-5dbb-4d7c-9cfb-64d6dccdb907 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.490428] env[65788]: DEBUG nova.network.neutron [-] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1220.562263] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3310e7f8-d659-4e5a-9f60-add4005cc78e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.573657] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29af787-71c2-46e4-834a-f455e685ac3e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.610698] env[65788]: DEBUG nova.compute.manager [req-b6848355-f091-4f64-98ad-75d213470bcf req-bcc2c13f-62b0-4998-9164-6d0a7d85a8ce service nova] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Detach interface failed, port_id=c703390b-d2a4-4b35-b51c-9b66c1714fb2, reason: Instance 152b3854-f2c6-471b-ac89-b32c51cfb762 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1220.721692] env[65788]: DEBUG oslo_concurrency.lockutils [req-a6276b13-c148-4e81-820c-b098c92144da req-a9c56e95-7169-4ccf-99d3-6cc7eed5a4ab service nova] Releasing lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1220.798228] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502348} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.798428] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1220.798644] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1220.798921] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b273d78d-a6e3-4ee4-8ad2-1cf8fb91c63e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.808621] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1220.808621] env[65788]: value = "task-4663355" [ 1220.808621] env[65788]: _type = "Task" [ 1220.808621] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.817688] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663355, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.995065] env[65788]: INFO nova.compute.manager [-] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Took 1.29 seconds to deallocate network for instance. [ 1220.997206] env[65788]: INFO nova.compute.manager [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Took 14.92 seconds to build instance. [ 1221.090360] env[65788]: DEBUG nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Successfully updated port: 114649fa-5d38-4198-8cf6-45044b37f249 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1221.207910] env[65788]: DEBUG nova.compute.manager [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-vif-plugged-114649fa-5d38-4198-8cf6-45044b37f249 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1221.207961] env[65788]: DEBUG oslo_concurrency.lockutils [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.208747] env[65788]: DEBUG oslo_concurrency.lockutils [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1221.209014] env[65788]: DEBUG oslo_concurrency.lockutils [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.209404] env[65788]: DEBUG nova.compute.manager [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] No waiting events found dispatching network-vif-plugged-114649fa-5d38-4198-8cf6-45044b37f249 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1221.209404] env[65788]: WARNING nova.compute.manager [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received unexpected event network-vif-plugged-114649fa-5d38-4198-8cf6-45044b37f249 for instance with vm_state active and task_state None. [ 1221.209595] env[65788]: DEBUG nova.compute.manager [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-changed-114649fa-5d38-4198-8cf6-45044b37f249 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1221.209595] env[65788]: DEBUG nova.compute.manager [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing instance network info cache due to event network-changed-114649fa-5d38-4198-8cf6-45044b37f249. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1221.209772] env[65788]: DEBUG oslo_concurrency.lockutils [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.210281] env[65788]: DEBUG oslo_concurrency.lockutils [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1221.210281] env[65788]: DEBUG nova.network.neutron [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Refreshing network info cache for port 114649fa-5d38-4198-8cf6-45044b37f249 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1221.320189] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663355, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083419} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.320547] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1221.321379] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33da8760-8136-4eb8-b106-dab8eba8403e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.347854] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1221.350994] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-780f7b2f-4e6d-490b-8436-f717d4a1323b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.372075] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1221.372075] env[65788]: value = "task-4663356" [ 1221.372075] env[65788]: _type = "Task" [ 1221.372075] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.385391] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663356, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.455615] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208e7503-d6cd-4820-9737-53cdb69ddf8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.464291] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fe4896-46c7-4482-9357-d6c4e3bfb746 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.498621] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614fef75-5597-49aa-a6b6-186f1271ce2d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.501523] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f8daa762-f3b3-4080-ac69-faa59d0fd8da tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.430s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1221.503015] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.508681] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9c0d7f-c1bd-412e-8928-74a1aa1cee6c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.525115] env[65788]: DEBUG nova.compute.provider_tree [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.593824] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.712343] env[65788]: WARNING neutronclient.v2_0.client [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.713186] env[65788]: WARNING openstack [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.713548] env[65788]: WARNING openstack [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.883482] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663356, 'name': ReconfigVM_Task, 'duration_secs': 0.30385} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.883884] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Reconfigured VM instance instance-00000078 to attach disk [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1221.884659] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53cfb087-cdbe-4652-9a2d-21571cb9df6f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.893180] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1221.893180] env[65788]: value = "task-4663357" [ 1221.893180] env[65788]: _type = "Task" [ 1221.893180] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.903017] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663357, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.904507] env[65788]: WARNING openstack [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.904932] env[65788]: WARNING openstack [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1221.979973] env[65788]: WARNING neutronclient.v2_0.client [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1221.980751] env[65788]: WARNING openstack [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1221.981170] env[65788]: WARNING openstack [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.030304] env[65788]: DEBUG nova.scheduler.client.report [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1222.070637] env[65788]: DEBUG nova.network.neutron [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Added VIF to instance network info cache for port 114649fa-5d38-4198-8cf6-45044b37f249. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3606}} [ 1222.071157] env[65788]: DEBUG nova.network.neutron [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c720474e-625c-4e3a-8654-dcfc4227f500", "address": "fa:16:3e:23:6d:d3", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc720474e-62", "ovs_interfaceid": "c720474e-625c-4e3a-8654-dcfc4227f500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "114649fa-5d38-4198-8cf6-45044b37f249", "address": "fa:16:3e:32:c8:97", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114649fa-5d", "ovs_interfaceid": "114649fa-5d38-4198-8cf6-45044b37f249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1222.337046] env[65788]: DEBUG nova.compute.manager [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Received event network-changed-f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1222.337046] env[65788]: DEBUG nova.compute.manager [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Refreshing instance network info cache due to event network-changed-f42a00fc-23eb-4df0-a834-72750f41e45f. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1222.337046] env[65788]: DEBUG oslo_concurrency.lockutils [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Acquiring lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.337046] env[65788]: DEBUG oslo_concurrency.lockutils [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Acquired lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.337458] env[65788]: DEBUG nova.network.neutron [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Refreshing network info cache for port f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1222.403542] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663357, 'name': Rename_Task, 'duration_secs': 0.154939} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.403813] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1222.404094] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cea64e0-2950-49a7-aaf8-bd2ed689e47a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.411853] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1222.411853] env[65788]: value = "task-4663358" [ 1222.411853] env[65788]: _type = "Task" [ 1222.411853] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.420799] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663358, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.536058] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.246s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.536470] env[65788]: WARNING neutronclient.v2_0.client [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.539133] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.036s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1222.539401] env[65788]: DEBUG nova.objects.instance [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid 152b3854-f2c6-471b-ac89-b32c51cfb762 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.572438] env[65788]: INFO nova.network.neutron [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating port c6074f79-c6f2-4186-b2c3-809f045737ed with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1222.575334] env[65788]: DEBUG oslo_concurrency.lockutils [req-be863e8c-120e-4753-8ae8-791077475629 req-be92d53c-ddb3-464d-a491-55e7d82c7395 service nova] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1222.576098] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.576348] env[65788]: DEBUG nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1222.840420] env[65788]: WARNING neutronclient.v2_0.client [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1222.841226] env[65788]: WARNING openstack [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1222.841583] env[65788]: WARNING openstack [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1222.922726] env[65788]: DEBUG oslo_vmware.api [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663358, 'name': PowerOnVM_Task, 'duration_secs': 0.465129} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.923045] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1222.923270] env[65788]: INFO nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1222.923651] env[65788]: DEBUG nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1222.924277] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714446d2-2f29-47d5-8182-700052d982a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.032647] env[65788]: WARNING openstack [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.033106] env[65788]: WARNING openstack [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.081078] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.081479] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.107054] env[65788]: WARNING neutronclient.v2_0.client [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.107723] env[65788]: WARNING openstack [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.108076] env[65788]: WARNING openstack [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.140555] env[65788]: WARNING nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1223.140820] env[65788]: WARNING nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1223.140985] env[65788]: WARNING nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1223.141163] env[65788]: WARNING nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] 114649fa-5d38-4198-8cf6-45044b37f249 already exists in list: port_ids containing: ['114649fa-5d38-4198-8cf6-45044b37f249']. ignoring it [ 1223.183141] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e927f5b4-46b4-4e95-96ce-f702e49b13ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.191304] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8602f0b-205b-4892-9c9b-7ebba3097ecc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.225605] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.226083] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.233131] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6381c2a5-816d-45e8-9864-2b3fec2b3235 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.241914] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390fb957-29ff-464a-800b-f3ab14265c5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.248643] env[65788]: DEBUG nova.network.neutron [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updated VIF entry in instance network info cache for port f42a00fc-23eb-4df0-a834-72750f41e45f. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1223.249028] env[65788]: DEBUG nova.network.neutron [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updating instance_info_cache with network_info: [{"id": "f42a00fc-23eb-4df0-a834-72750f41e45f", "address": "fa:16:3e:59:ac:ac", "network": {"id": "b66b29a1-3158-48a6-93d0-5c6d304c9990", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-271434673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7aa1dec53aa4436daa0f3223b92f9b79", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "35fcdc55-dc29-451b-ad56-3a03b044dc81", "external-id": "nsx-vlan-transportzone-552", "segmentation_id": 552, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf42a00fc-23", "ovs_interfaceid": "f42a00fc-23eb-4df0-a834-72750f41e45f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.261690] env[65788]: DEBUG nova.compute.provider_tree [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1223.313030] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.313796] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.314222] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.416435] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.416902] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.451021] env[65788]: INFO nova.compute.manager [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Took 15.10 seconds to build instance. [ 1223.483775] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.484490] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.484930] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.602399] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.602808] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.692610] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1223.693321] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1223.693663] env[65788]: WARNING openstack [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1223.752297] env[65788]: DEBUG oslo_concurrency.lockutils [req-4e91308a-1315-48bd-9c3b-42e3d404548d req-85bf140b-a331-427f-a9d2-f1ddfaed7665 service nova] Releasing lock "refresh_cache-ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1223.765014] env[65788]: DEBUG nova.scheduler.client.report [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1223.793166] env[65788]: DEBUG nova.network.neutron [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c720474e-625c-4e3a-8654-dcfc4227f500", "address": "fa:16:3e:23:6d:d3", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc720474e-62", "ovs_interfaceid": "c720474e-625c-4e3a-8654-dcfc4227f500", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "114649fa-5d38-4198-8cf6-45044b37f249", "address": "fa:16:3e:32:c8:97", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114649fa-5d", "ovs_interfaceid": "114649fa-5d38-4198-8cf6-45044b37f249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1223.953204] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fc486e9b-60ad-4aa2-86da-9a78d7b04f08 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "222d1456-baad-42ee-bdef-020fb66ebe41" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.605s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.105234] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.105234] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.105234] env[65788]: DEBUG nova.network.neutron [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1224.270025] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.295539] env[65788]: INFO nova.scheduler.client.report [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance 152b3854-f2c6-471b-ac89-b32c51cfb762 [ 1224.297191] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.297741] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.297901] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.299739] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7640e8-25bc-46a0-b647-e9bf4963a721 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.322198] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1224.322474] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1224.322662] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1224.322857] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1224.323284] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1224.323284] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1224.323433] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1224.323526] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1224.323690] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1224.324235] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1224.324235] env[65788]: DEBUG nova.virt.hardware [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1224.330664] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfiguring VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1224.331084] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b4d3982-2282-4885-8cfb-ed1b382aab43 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.354026] env[65788]: DEBUG oslo_vmware.api [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1224.354026] env[65788]: value = "task-4663359" [ 1224.354026] env[65788]: _type = "Task" [ 1224.354026] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.363387] env[65788]: DEBUG oslo_vmware.api [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663359, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.377031] env[65788]: DEBUG nova.compute.manager [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Received event network-changed-fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1224.377322] env[65788]: DEBUG nova.compute.manager [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Refreshing instance network info cache due to event network-changed-fff8c844-a1fe-4657-9ab0-99fa3a35d346. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1224.377644] env[65788]: DEBUG oslo_concurrency.lockutils [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Acquiring lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.377815] env[65788]: DEBUG oslo_concurrency.lockutils [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Acquired lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.377980] env[65788]: DEBUG nova.network.neutron [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Refreshing network info cache for port fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1224.401469] env[65788]: DEBUG nova.compute.manager [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-vif-plugged-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1224.401469] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1224.401638] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Lock "bdd09116-31de-491d-b129-c117f898881e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1224.401729] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Lock "bdd09116-31de-491d-b129-c117f898881e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.401901] env[65788]: DEBUG nova.compute.manager [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] No waiting events found dispatching network-vif-plugged-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1224.402078] env[65788]: WARNING nova.compute.manager [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received unexpected event network-vif-plugged-c6074f79-c6f2-4186-b2c3-809f045737ed for instance with vm_state shelved_offloaded and task_state spawning. [ 1224.402233] env[65788]: DEBUG nova.compute.manager [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1224.402404] env[65788]: DEBUG nova.compute.manager [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing instance network info cache due to event network-changed-c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1224.402651] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Acquiring lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.609140] env[65788]: WARNING neutronclient.v2_0.client [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.610098] env[65788]: WARNING openstack [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.610376] env[65788]: WARNING openstack [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.736372] env[65788]: WARNING openstack [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.736866] env[65788]: WARNING openstack [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.806186] env[65788]: WARNING neutronclient.v2_0.client [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.806912] env[65788]: WARNING openstack [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.807369] env[65788]: WARNING openstack [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.818945] env[65788]: DEBUG oslo_concurrency.lockutils [None req-115aeae1-4486-461f-a386-325f383557e5 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "152b3854-f2c6-471b-ac89-b32c51cfb762" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.772s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.864550] env[65788]: DEBUG oslo_vmware.api [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663359, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.881023] env[65788]: WARNING neutronclient.v2_0.client [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1224.881973] env[65788]: WARNING openstack [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1224.882806] env[65788]: WARNING openstack [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1224.939813] env[65788]: DEBUG nova.network.neutron [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1225.010244] env[65788]: WARNING openstack [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.010742] env[65788]: WARNING openstack [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.075780] env[65788]: WARNING neutronclient.v2_0.client [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1225.076501] env[65788]: WARNING openstack [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.076927] env[65788]: WARNING openstack [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.164228] env[65788]: DEBUG nova.network.neutron [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updated VIF entry in instance network info cache for port fff8c844-a1fe-4657-9ab0-99fa3a35d346. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1225.164865] env[65788]: DEBUG nova.network.neutron [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updating instance_info_cache with network_info: [{"id": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "address": "fa:16:3e:f7:50:eb", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfff8c844-a1", "ovs_interfaceid": "fff8c844-a1fe-4657-9ab0-99fa3a35d346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1225.363536] env[65788]: DEBUG oslo_vmware.api [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663359, 'name': ReconfigVM_Task, 'duration_secs': 0.617909} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.364048] env[65788]: WARNING neutronclient.v2_0.client [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1225.364331] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.364543] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfigured VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1225.443181] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.445976] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Acquired lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1225.446764] env[65788]: DEBUG nova.network.neutron [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Refreshing network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1225.473049] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5d12ab7008502b05bb0bb86374ef3240',container_format='bare',created_at=2025-11-21T13:56:01Z,direct_url=,disk_format='vmdk',id=e741c886-f6dd-49f9-b52f-b3ee9f0d3301,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-169206401-shelved',owner='94977b75768444138fb914cabddc57c5',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2025-11-21T13:56:15Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1225.473049] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1225.473263] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1225.473381] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1225.473529] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1225.473676] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1225.473885] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1225.474056] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1225.474236] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1225.474393] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1225.474580] env[65788]: DEBUG nova.virt.hardware [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1225.475999] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d594e2c8-22c7-40d2-bc14-4db6aa89fa55 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.486253] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff47f70-15ab-46aa-9ff3-900bc4808f97 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.501971] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:d9:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a0a76279-3c11-4bef-b124-2a2ee13fa377', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6074f79-c6f2-4186-b2c3-809f045737ed', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1225.509832] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1225.510645] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1225.510944] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0c2c287-3577-4edb-b316-7d7ddadac022 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.533539] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1225.533539] env[65788]: value = "task-4663360" [ 1225.533539] env[65788]: _type = "Task" [ 1225.533539] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.546236] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663360, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.668414] env[65788]: DEBUG oslo_concurrency.lockutils [req-a419c057-be45-4ed9-8ddd-244530217ccc req-4d0d02d0-1540-41ad-aaad-4520a2062161 service nova] Releasing lock "refresh_cache-222d1456-baad-42ee-bdef-020fb66ebe41" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.869541] env[65788]: DEBUG oslo_concurrency.lockutils [None req-59c9da07-8b6c-41a3-9199-1df5ae1fccf5 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-114649fa-5d38-4198-8cf6-45044b37f249" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.523s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.953539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "d15953f8-f74a-422a-867d-fa835bb479ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.953805] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "d15953f8-f74a-422a-867d-fa835bb479ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.954090] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "d15953f8-f74a-422a-867d-fa835bb479ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.954290] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "d15953f8-f74a-422a-867d-fa835bb479ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.954461] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "d15953f8-f74a-422a-867d-fa835bb479ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1225.956982] env[65788]: WARNING neutronclient.v2_0.client [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1225.957643] env[65788]: WARNING openstack [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1225.957989] env[65788]: WARNING openstack [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1225.966670] env[65788]: INFO nova.compute.manager [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Terminating instance [ 1226.044344] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663360, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.088471] env[65788]: WARNING openstack [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.088932] env[65788]: WARNING openstack [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.164055] env[65788]: WARNING neutronclient.v2_0.client [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1226.164823] env[65788]: WARNING openstack [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1226.165379] env[65788]: WARNING openstack [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1226.260044] env[65788]: DEBUG nova.network.neutron [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updated VIF entry in instance network info cache for port c6074f79-c6f2-4186-b2c3-809f045737ed. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1226.260619] env[65788]: DEBUG nova.network.neutron [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [{"id": "c6074f79-c6f2-4186-b2c3-809f045737ed", "address": "fa:16:3e:01:d9:0e", "network": {"id": "5eea60cb-5c11-4331-86c7-190d4b6dd829", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-458650503-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "94977b75768444138fb914cabddc57c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a0a76279-3c11-4bef-b124-2a2ee13fa377", "external-id": "nsx-vlan-transportzone-738", "segmentation_id": 738, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6074f79-c6", "ovs_interfaceid": "c6074f79-c6f2-4186-b2c3-809f045737ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1226.471669] env[65788]: DEBUG nova.compute.manager [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1226.471904] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1226.472967] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de1d52f-0acc-4cb6-b211-b1fc1e76e794 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.481493] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1226.481739] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f07b9d8-1fbb-42f1-8379-85cbfd344e01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.488432] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1226.488432] env[65788]: value = "task-4663361" [ 1226.488432] env[65788]: _type = "Task" [ 1226.488432] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.497986] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.548920] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663360, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.764076] env[65788]: DEBUG oslo_concurrency.lockutils [req-6f390aaa-bcf2-4809-94e0-b12bda9b1d15 req-58868911-a16d-4a6b-8038-e010b3e26ad4 service nova] Releasing lock "refresh_cache-bdd09116-31de-491d-b129-c117f898881e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.999656] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.047440] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663360, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.500445] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663361, 'name': PowerOffVM_Task, 'duration_secs': 0.81198} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.500707] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1227.500917] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1227.501147] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f866f2d-6190-47e8-a05c-371d6d3b09bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.544784] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663360, 'name': CreateVM_Task, 'duration_secs': 1.56566} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.544974] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1227.545508] env[65788]: WARNING neutronclient.v2_0.client [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1227.545884] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1227.546086] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1227.546474] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1227.546725] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74947cd1-cd59-4209-9419-be759f7af292 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.552614] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1227.552614] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523f87a5-32e6-1931-447d-e7739c91efc6" [ 1227.552614] env[65788]: _type = "Task" [ 1227.552614] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.563518] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]523f87a5-32e6-1931-447d-e7739c91efc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.575846] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1227.576143] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1227.576361] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore2] d15953f8-f74a-422a-867d-fa835bb479ff {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1227.576654] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cee5d746-3761-48a5-9c31-1688acb1684b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.583838] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1227.583838] env[65788]: value = "task-4663363" [ 1227.583838] env[65788]: _type = "Task" [ 1227.583838] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.593102] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663363, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.703303] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-f8c57215-ade6-424b-be8c-075998a764af-c720474e-625c-4e3a-8654-dcfc4227f500" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.703651] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-c720474e-625c-4e3a-8654-dcfc4227f500" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1228.064835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1228.065250] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Processing image e741c886-f6dd-49f9-b52f-b3ee9f0d3301 {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1228.065318] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.065466] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1228.065648] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1228.065919] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dace726-bcfb-411c-8f42-63a5b792cc99 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.078709] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1228.078892] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1228.079663] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-411cd0db-b75f-4bfb-8c96-80cd869f533b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.085751] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1228.085751] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52631a6a-5844-8752-585e-f75a73cd995a" [ 1228.085751] env[65788]: _type = "Task" [ 1228.085751] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.097669] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52631a6a-5844-8752-585e-f75a73cd995a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.100652] env[65788]: DEBUG oslo_vmware.api [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663363, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14042} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.100925] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1228.101089] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1228.101264] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1228.101438] env[65788]: INFO nova.compute.manager [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1228.101683] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1228.101880] env[65788]: DEBUG nova.compute.manager [-] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1228.101978] env[65788]: DEBUG nova.network.neutron [-] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1228.102235] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.102761] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1228.103015] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1228.139663] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.207436] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.207436] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1228.208314] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a741f6f-ec00-4f36-af1b-46aa6e644037 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.227797] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8e9649-a243-4927-b669-ad8b29c26220 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.253905] env[65788]: WARNING neutronclient.v2_0.client [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1228.259588] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfiguring VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1228.259930] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3254673-88bd-48ca-9228-fb96155cc312 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.282082] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1228.282082] env[65788]: value = "task-4663364" [ 1228.282082] env[65788]: _type = "Task" [ 1228.282082] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.291799] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.377860] env[65788]: DEBUG nova.compute.manager [req-412092ed-ce45-4617-9b8a-5201b858e11a req-a454880b-9717-4ff9-8f42-29c298a95dca service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Received event network-vif-deleted-7bd8a945-5b5b-43be-8383-f403522245b3 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1228.378073] env[65788]: INFO nova.compute.manager [req-412092ed-ce45-4617-9b8a-5201b858e11a req-a454880b-9717-4ff9-8f42-29c298a95dca service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Neutron deleted interface 7bd8a945-5b5b-43be-8383-f403522245b3; detaching it from the instance and deleting it from the info cache [ 1228.378246] env[65788]: DEBUG nova.network.neutron [req-412092ed-ce45-4617-9b8a-5201b858e11a req-a454880b-9717-4ff9-8f42-29c298a95dca service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1228.600139] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Preparing fetch location {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1228.600402] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Fetch image to [datastore1] OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f/OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f.vmdk {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1228.600562] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Downloading stream optimized image e741c886-f6dd-49f9-b52f-b3ee9f0d3301 to [datastore1] OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f/OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f.vmdk on the data store datastore1 as vApp {{(pid=65788) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1228.600735] env[65788]: DEBUG nova.virt.vmwareapi.images [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Downloading image file data e741c886-f6dd-49f9-b52f-b3ee9f0d3301 to the ESX as VM named 'OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f' {{(pid=65788) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1228.691014] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1228.691014] env[65788]: value = "resgroup-9" [ 1228.691014] env[65788]: _type = "ResourcePool" [ 1228.691014] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1228.691014] env[65788]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-c36634fc-7615-48e0-8a3e-7108238114e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.711628] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease: (returnval){ [ 1228.711628] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1228.711628] env[65788]: _type = "HttpNfcLease" [ 1228.711628] env[65788]: } obtained for vApp import into resource pool (val){ [ 1228.711628] env[65788]: value = "resgroup-9" [ 1228.711628] env[65788]: _type = "ResourcePool" [ 1228.711628] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1228.712186] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the lease: (returnval){ [ 1228.712186] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1228.712186] env[65788]: _type = "HttpNfcLease" [ 1228.712186] env[65788]: } to be ready. {{(pid=65788) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1228.719668] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1228.719668] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1228.719668] env[65788]: _type = "HttpNfcLease" [ 1228.719668] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1228.793369] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.850260] env[65788]: DEBUG nova.network.neutron [-] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1228.881269] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4499ca74-2c01-4e1f-9625-25c39c8b8864 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.892745] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8d7036-9076-495f-99a3-d227bd96479d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.928756] env[65788]: DEBUG nova.compute.manager [req-412092ed-ce45-4617-9b8a-5201b858e11a req-a454880b-9717-4ff9-8f42-29c298a95dca service nova] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Detach interface failed, port_id=7bd8a945-5b5b-43be-8383-f403522245b3, reason: Instance d15953f8-f74a-422a-867d-fa835bb479ff could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1229.220968] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1229.220968] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1229.220968] env[65788]: _type = "HttpNfcLease" [ 1229.220968] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1229.293714] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.353604] env[65788]: INFO nova.compute.manager [-] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Took 1.25 seconds to deallocate network for instance. [ 1229.720188] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1229.720188] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1229.720188] env[65788]: _type = "HttpNfcLease" [ 1229.720188] env[65788]: } is initializing. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1229.795768] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.861333] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1229.861691] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1229.861956] env[65788]: DEBUG nova.objects.instance [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid d15953f8-f74a-422a-867d-fa835bb479ff {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1230.222198] env[65788]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1230.222198] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1230.222198] env[65788]: _type = "HttpNfcLease" [ 1230.222198] env[65788]: } is ready. {{(pid=65788) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1230.222664] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1230.222664] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52779d60-e357-40d8-4ae7-ad65721022c8" [ 1230.222664] env[65788]: _type = "HttpNfcLease" [ 1230.222664] env[65788]: }. {{(pid=65788) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1230.223235] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ccc885-143d-4897-aec9-af55ee262bad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.231220] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52935c2e-da3e-41e4-c3aa-d4a227290f4d/disk-0.vmdk from lease info. {{(pid=65788) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1230.231409] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52935c2e-da3e-41e4-c3aa-d4a227290f4d/disk-0.vmdk. {{(pid=65788) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1230.297677] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-96c8580b-237c-4555-b8ed-b0869a9840da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.299587] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.485918] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfacdcf4-e106-4186-8751-efd26a028eb7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.494434] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24903918-97d6-4294-b30f-10a9bfad0c8c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.532330] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a60f640-a175-4abe-bab8-bfe932d4762e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.541743] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f179f2e-0635-41fd-9bdc-2306b3530316 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.560607] env[65788]: DEBUG nova.compute.provider_tree [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.795893] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.066766] env[65788]: DEBUG nova.scheduler.client.report [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1231.297166] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.573919] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.712s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.597704] env[65788]: INFO nova.scheduler.client.report [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance d15953f8-f74a-422a-867d-fa835bb479ff [ 1231.696620] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Completed reading data from the image iterator. {{(pid=65788) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1231.696858] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52935c2e-da3e-41e4-c3aa-d4a227290f4d/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1231.697847] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce147e76-6eda-4ec4-a644-cb98f1b15a42 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.705197] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52935c2e-da3e-41e4-c3aa-d4a227290f4d/disk-0.vmdk is in state: ready. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1231.705413] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52935c2e-da3e-41e4-c3aa-d4a227290f4d/disk-0.vmdk. {{(pid=65788) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1231.705626] env[65788]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-8a440376-bb61-4660-ab5e-e7a526bf379a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.799088] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.909866] env[65788]: DEBUG oslo_vmware.rw_handles [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52935c2e-da3e-41e4-c3aa-d4a227290f4d/disk-0.vmdk. {{(pid=65788) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1231.910160] env[65788]: INFO nova.virt.vmwareapi.images [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Downloaded image file data e741c886-f6dd-49f9-b52f-b3ee9f0d3301 [ 1231.911172] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9718126d-a9a4-46f2-8447-ae4f2d885ef7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.928372] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e78fabcf-7742-48d7-a4f2-9d6fea1d7986 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.022025] env[65788]: INFO nova.virt.vmwareapi.images [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] The imported VM was unregistered [ 1232.024980] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Caching image {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1232.025240] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Creating directory with path [datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301 {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1232.025550] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e3f2b0a-e14d-4589-a693-69a80da9e2f6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.040946] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Created directory with path [datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301 {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1232.041170] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f/OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f.vmdk to [datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk. {{(pid=65788) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1232.041443] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f7c00747-aa53-4aba-9189-7659ab47999f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.049087] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1232.049087] env[65788]: value = "task-4663367" [ 1232.049087] env[65788]: _type = "Task" [ 1232.049087] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.058050] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663367, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.106118] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1ef09e60-5d4c-40db-8eb9-654376d59a69 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "d15953f8-f74a-422a-867d-fa835bb479ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.152s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1232.299874] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.560221] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663367, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.800020] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.061819] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663367, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.123978] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.124297] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.301932] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.563440] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663367, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.626571] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1233.801790] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.067606] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663367, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.158237] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1234.158566] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1234.160276] env[65788]: INFO nova.compute.claims [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1234.301811] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.563586] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663367, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.444964} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.563877] env[65788]: INFO nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f/OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f.vmdk to [datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk. [ 1234.564064] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Cleaning up location [datastore1] OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1234.564232] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_7aa15a4e-72c2-4aa4-a04c-e1df0328ec0f {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1234.564497] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-121e6a52-8f1e-4f99-8700-996b25f403e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.571145] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1234.571145] env[65788]: value = "task-4663368" [ 1234.571145] env[65788]: _type = "Task" [ 1234.571145] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.579439] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663368, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.801690] env[65788]: DEBUG oslo_vmware.api [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663364, 'name': ReconfigVM_Task, 'duration_secs': 6.196018} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.802042] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1234.802290] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfigured VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1234.802767] env[65788]: WARNING neutronclient.v2_0.client [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1234.803145] env[65788]: WARNING neutronclient.v2_0.client [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1234.803769] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1234.804165] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1235.082981] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663368, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037036} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.083555] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.083555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1235.083740] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk to [datastore1] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1235.084029] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9df3182-cbf6-40f0-b18b-97d0998c8649 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.092453] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1235.092453] env[65788]: value = "task-4663369" [ 1235.092453] env[65788]: _type = "Task" [ 1235.092453] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.103414] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.121995] env[65788]: DEBUG nova.compute.manager [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-vif-deleted-c720474e-625c-4e3a-8654-dcfc4227f500 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1235.122370] env[65788]: INFO nova.compute.manager [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Neutron deleted interface c720474e-625c-4e3a-8654-dcfc4227f500; detaching it from the instance and deleting it from the info cache [ 1235.122518] env[65788]: DEBUG nova.network.neutron [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "114649fa-5d38-4198-8cf6-45044b37f249", "address": "fa:16:3e:32:c8:97", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114649fa-5d", "ovs_interfaceid": "114649fa-5d38-4198-8cf6-45044b37f249", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1235.285182] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1041770d-8446-4490-80ca-334c52a8a6ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.293755] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58b8a49-81f0-4fe5-bb0a-e9ed066fa8cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.325615] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57641d20-9e80-47cb-a2db-8928f6f28303 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.334056] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1160b032-7aa4-4077-b25e-bf02b5b95412 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.348567] env[65788]: DEBUG nova.compute.provider_tree [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.603611] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.626068] env[65788]: DEBUG oslo_concurrency.lockutils [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.626281] env[65788]: DEBUG oslo_concurrency.lockutils [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] Acquired lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1235.627537] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd994a1-2b37-4aed-91ff-2b6becf197d9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.648860] env[65788]: DEBUG oslo_concurrency.lockutils [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] Releasing lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1235.649147] env[65788]: WARNING nova.compute.manager [req-83dbec43-a719-42bf-a674-b3becd73fc08 req-01f856d4-9b4d-4f07-93e6-719d523119f9 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Detach interface failed, port_id=c720474e-625c-4e3a-8654-dcfc4227f500, reason: No device with interface-id c720474e-625c-4e3a-8654-dcfc4227f500 exists on VM: nova.exception.NotFound: No device with interface-id c720474e-625c-4e3a-8654-dcfc4227f500 exists on VM [ 1235.852121] env[65788]: DEBUG nova.scheduler.client.report [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1236.104957] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.105174] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1236.105378] env[65788]: DEBUG nova.network.neutron [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1236.106992] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.255473] env[65788]: DEBUG nova.compute.manager [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-vif-deleted-114649fa-5d38-4198-8cf6-45044b37f249 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1236.255593] env[65788]: INFO nova.compute.manager [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Neutron deleted interface 114649fa-5d38-4198-8cf6-45044b37f249; detaching it from the instance and deleting it from the info cache [ 1236.256775] env[65788]: DEBUG nova.network.neutron [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1236.357517] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.358110] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1236.604762] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.608789] env[65788]: WARNING neutronclient.v2_0.client [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.609493] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.609838] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.743672] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.744129] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.760173] env[65788]: DEBUG oslo_concurrency.lockutils [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.760420] env[65788]: DEBUG oslo_concurrency.lockutils [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Acquired lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1236.761957] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49baae65-a78f-49f9-964a-9563b8387195 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.787594] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012f930f-fc80-4b3d-a50a-a8834da2c1d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.812218] env[65788]: WARNING neutronclient.v2_0.client [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.818251] env[65788]: DEBUG nova.virt.vmwareapi.vmops [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfiguring VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1236.818626] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d87d7ff5-7107-4e4c-b961-a74a40e4db44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.833692] env[65788]: WARNING neutronclient.v2_0.client [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.834384] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.834765] env[65788]: WARNING openstack [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.842575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.851564] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Waiting for the task: (returnval){ [ 1236.851564] env[65788]: value = "task-4663370" [ 1236.851564] env[65788]: _type = "Task" [ 1236.851564] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.863428] env[65788]: DEBUG nova.compute.utils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1236.864913] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.865439] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1236.865704] env[65788]: DEBUG nova.network.neutron [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1236.866053] env[65788]: WARNING neutronclient.v2_0.client [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.866366] env[65788]: WARNING neutronclient.v2_0.client [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1236.867023] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1236.867426] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1236.917384] env[65788]: DEBUG nova.policy [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1236.930591] env[65788]: INFO nova.network.neutron [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Port 114649fa-5d38-4198-8cf6-45044b37f249 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1236.931067] env[65788]: DEBUG nova.network.neutron [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [{"id": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "address": "fa:16:3e:9a:90:23", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfda83fac-56", "ovs_interfaceid": "fda83fac-56b6-4ab0-824a-7af792b7c8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1237.106330] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.205342] env[65788]: DEBUG nova.network.neutron [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Successfully created port: b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1237.364018] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.366839] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1237.433654] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1237.606981] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.863032] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.938057] env[65788]: DEBUG oslo_concurrency.lockutils [None req-336e0c12-0a97-4ec0-98ea-5bed9950c180 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-f8c57215-ade6-424b-be8c-075998a764af-c720474e-625c-4e3a-8654-dcfc4227f500" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.234s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.106834] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663369, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.6451} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.107139] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e741c886-f6dd-49f9-b52f-b3ee9f0d3301/e741c886-f6dd-49f9-b52f-b3ee9f0d3301.vmdk to [datastore1] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1238.107946] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17414fab-bc69-49c4-a7f0-c2ea1064e849 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.131465] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1238.131743] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34359338-2360-478d-b366-0520b6f08399 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.152445] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1238.152445] env[65788]: value = "task-4663371" [ 1238.152445] env[65788]: _type = "Task" [ 1238.152445] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.162546] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.364120] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.377614] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1238.404506] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1238.404827] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1238.405100] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1238.405299] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1238.405444] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1238.405584] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1238.405841] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1238.406067] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1238.406252] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1238.406419] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1238.406591] env[65788]: DEBUG nova.virt.hardware [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1238.407481] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e745a7e-9014-4646-99c7-0da35f731356 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.415913] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8314eb01-08c9-474c-9477-9ef5b1968b28 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.664330] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663371, 'name': ReconfigVM_Task, 'duration_secs': 0.325753} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.665031] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Reconfigured VM instance instance-0000006f to attach disk [datastore1] bdd09116-31de-491d-b129-c117f898881e/bdd09116-31de-491d-b129-c117f898881e.vmdk or device None with type streamOptimized {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1238.665252] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7abd4707-99b3-4171-821f-76f999ab133e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.674069] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1238.674069] env[65788]: value = "task-4663372" [ 1238.674069] env[65788]: _type = "Task" [ 1238.674069] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.682873] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663372, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.697664] env[65788]: DEBUG nova.compute.manager [req-b6b0b672-92ed-45a8-aa1a-893aafed2a02 req-ead94fe7-be7d-44fe-95c2-368449b0b6c0 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Received event network-vif-plugged-b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1238.698011] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6b0b672-92ed-45a8-aa1a-893aafed2a02 req-ead94fe7-be7d-44fe-95c2-368449b0b6c0 service nova] Acquiring lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1238.698410] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6b0b672-92ed-45a8-aa1a-893aafed2a02 req-ead94fe7-be7d-44fe-95c2-368449b0b6c0 service nova] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1238.698410] env[65788]: DEBUG oslo_concurrency.lockutils [req-b6b0b672-92ed-45a8-aa1a-893aafed2a02 req-ead94fe7-be7d-44fe-95c2-368449b0b6c0 service nova] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.698625] env[65788]: DEBUG nova.compute.manager [req-b6b0b672-92ed-45a8-aa1a-893aafed2a02 req-ead94fe7-be7d-44fe-95c2-368449b0b6c0 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] No waiting events found dispatching network-vif-plugged-b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1238.698842] env[65788]: WARNING nova.compute.manager [req-b6b0b672-92ed-45a8-aa1a-893aafed2a02 req-ead94fe7-be7d-44fe-95c2-368449b0b6c0 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Received unexpected event network-vif-plugged-b9563bd8-34ff-48e6-9804-3643714bb99a for instance with vm_state building and task_state spawning. [ 1238.826044] env[65788]: DEBUG nova.network.neutron [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Successfully updated port: b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1238.866279] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.183900] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663372, 'name': Rename_Task, 'duration_secs': 0.153399} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.184207] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1239.184473] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdf512b7-0997-4e6f-9ba3-3974fece4b3b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.191667] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1239.191667] env[65788]: value = "task-4663373" [ 1239.191667] env[65788]: _type = "Task" [ 1239.191667] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.201242] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.329796] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-67934662-1b06-47b4-9fdf-d8f6ba17b86c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.330068] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-67934662-1b06-47b4-9fdf-d8f6ba17b86c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1239.330241] env[65788]: DEBUG nova.network.neutron [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1239.366494] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.701908] env[65788]: DEBUG oslo_vmware.api [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663373, 'name': PowerOnVM_Task, 'duration_secs': 0.466367} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.702305] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1239.808502] env[65788]: DEBUG nova.compute.manager [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1239.809559] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ddff05-a392-45f7-80f7-7d63591c2b52 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.833177] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.833605] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.867948] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.872284] env[65788]: DEBUG nova.network.neutron [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1239.891950] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.892457] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1239.955521] env[65788]: WARNING neutronclient.v2_0.client [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1239.956228] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1239.956577] env[65788]: WARNING openstack [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1240.042072] env[65788]: DEBUG nova.network.neutron [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Updating instance_info_cache with network_info: [{"id": "b9563bd8-34ff-48e6-9804-3643714bb99a", "address": "fa:16:3e:29:80:87", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9563bd8-34", "ovs_interfaceid": "b9563bd8-34ff-48e6-9804-3643714bb99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1240.329563] env[65788]: DEBUG oslo_concurrency.lockutils [None req-8262f9be-a6db-4ed1-8da9-f278d0fe6d13 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 22.068s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.366970] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.544612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-67934662-1b06-47b4-9fdf-d8f6ba17b86c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1240.545034] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Instance network_info: |[{"id": "b9563bd8-34ff-48e6-9804-3643714bb99a", "address": "fa:16:3e:29:80:87", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9563bd8-34", "ovs_interfaceid": "b9563bd8-34ff-48e6-9804-3643714bb99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1240.545620] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:80:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9563bd8-34ff-48e6-9804-3643714bb99a', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1240.553251] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1240.553477] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1240.553704] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5c7f165-7576-4b55-bb57-69b846ce07ee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.574468] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1240.574468] env[65788]: value = "task-4663374" [ 1240.574468] env[65788]: _type = "Task" [ 1240.574468] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.582600] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663374, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.727046] env[65788]: DEBUG nova.compute.manager [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Received event network-changed-b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1240.727356] env[65788]: DEBUG nova.compute.manager [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Refreshing instance network info cache due to event network-changed-b9563bd8-34ff-48e6-9804-3643714bb99a. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1240.727573] env[65788]: DEBUG oslo_concurrency.lockutils [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Acquiring lock "refresh_cache-67934662-1b06-47b4-9fdf-d8f6ba17b86c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.727945] env[65788]: DEBUG oslo_concurrency.lockutils [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Acquired lock "refresh_cache-67934662-1b06-47b4-9fdf-d8f6ba17b86c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1240.728018] env[65788]: DEBUG nova.network.neutron [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Refreshing network info cache for port b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1240.867949] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.084548] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663374, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.231512] env[65788]: WARNING neutronclient.v2_0.client [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.232262] env[65788]: WARNING openstack [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.232640] env[65788]: WARNING openstack [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.326536] env[65788]: WARNING openstack [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.326964] env[65788]: WARNING openstack [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.368020] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.388245] env[65788]: WARNING neutronclient.v2_0.client [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.388935] env[65788]: WARNING openstack [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1241.389334] env[65788]: WARNING openstack [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1241.469455] env[65788]: DEBUG nova.network.neutron [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Updated VIF entry in instance network info cache for port b9563bd8-34ff-48e6-9804-3643714bb99a. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1241.469821] env[65788]: DEBUG nova.network.neutron [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Updating instance_info_cache with network_info: [{"id": "b9563bd8-34ff-48e6-9804-3643714bb99a", "address": "fa:16:3e:29:80:87", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9563bd8-34", "ovs_interfaceid": "b9563bd8-34ff-48e6-9804-3643714bb99a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1241.585415] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663374, 'name': CreateVM_Task, 'duration_secs': 0.544278} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.585622] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1241.586211] env[65788]: WARNING neutronclient.v2_0.client [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1241.586586] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.586737] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1241.587095] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1241.587363] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f985679-0eec-40bd-af94-3ff724e1b699 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.592340] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1241.592340] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52792f62-7745-09be-4c93-eb4bad5c8bd5" [ 1241.592340] env[65788]: _type = "Task" [ 1241.592340] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.600342] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52792f62-7745-09be-4c93-eb4bad5c8bd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.870129] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.972279] env[65788]: DEBUG oslo_concurrency.lockutils [req-f09bbd66-187c-429c-b23f-7245f98b5813 req-5e8884df-5fda-4681-8e4e-1a7964fa06e9 service nova] Releasing lock "refresh_cache-67934662-1b06-47b4-9fdf-d8f6ba17b86c" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.104589] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52792f62-7745-09be-4c93-eb4bad5c8bd5, 'name': SearchDatastore_Task, 'duration_secs': 0.010847} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.104954] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.105245] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1242.105489] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1242.105644] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1242.105856] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1242.106150] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6dd24ce1-ff6c-4175-91b9-6218dd54a7ff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.115614] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1242.115875] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1242.116666] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6eb9c4d2-5c34-44b2-bb3a-693a8b9614e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.122671] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1242.122671] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c732e-528b-6155-3974-feced3ac2eb9" [ 1242.122671] env[65788]: _type = "Task" [ 1242.122671] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.132744] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c732e-528b-6155-3974-feced3ac2eb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.370048] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.633895] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521c732e-528b-6155-3974-feced3ac2eb9, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.634733] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efd2b073-4c36-43bc-9252-042801d3890a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.640513] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1242.640513] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525acf76-60e6-9f83-e586-2afb756ec39f" [ 1242.640513] env[65788]: _type = "Task" [ 1242.640513] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.649106] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525acf76-60e6-9f83-e586-2afb756ec39f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.869108] env[65788]: DEBUG oslo_vmware.api [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Task: {'id': task-4663370, 'name': ReconfigVM_Task, 'duration_secs': 5.844574} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.869357] env[65788]: DEBUG oslo_concurrency.lockutils [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] Releasing lock "f8c57215-ade6-424b-be8c-075998a764af" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.869557] env[65788]: DEBUG nova.virt.vmwareapi.vmops [req-2156c05d-5152-457f-b6ac-fa881101fffd req-e6a690c1-5627-44db-b06f-c86eed48c240 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Reconfigured VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1242.870150] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "f8c57215-ade6-424b-be8c-075998a764af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.028s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.870437] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "f8c57215-ade6-424b-be8c-075998a764af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.870560] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.870724] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "f8c57215-ade6-424b-be8c-075998a764af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.872726] env[65788]: INFO nova.compute.manager [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Terminating instance [ 1243.151610] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525acf76-60e6-9f83-e586-2afb756ec39f, 'name': SearchDatastore_Task, 'duration_secs': 0.009121} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.151887] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1243.152163] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 67934662-1b06-47b4-9fdf-d8f6ba17b86c/67934662-1b06-47b4-9fdf-d8f6ba17b86c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1243.152434] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-adec3ed0-d095-45e1-92d4-58f9bb0c381d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.159385] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1243.159385] env[65788]: value = "task-4663375" [ 1243.159385] env[65788]: _type = "Task" [ 1243.159385] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.168193] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663375, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.376404] env[65788]: DEBUG nova.compute.manager [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1243.376649] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1243.377660] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9031b168-b70f-441b-aceb-46a277967cb6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.386120] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1243.386350] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d657b20-2ee7-4174-8a98-9770cba9525c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.393263] env[65788]: DEBUG oslo_vmware.api [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1243.393263] env[65788]: value = "task-4663376" [ 1243.393263] env[65788]: _type = "Task" [ 1243.393263] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.402130] env[65788]: DEBUG oslo_vmware.api [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663376, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.671030] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663375, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.904175] env[65788]: DEBUG oslo_vmware.api [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663376, 'name': PowerOffVM_Task, 'duration_secs': 0.49781} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.904550] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1243.904550] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1243.904768] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51fa0392-a845-4347-8978-506b47e10fe9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.976821] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1243.977064] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1243.977220] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleting the datastore file [datastore2] f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1243.977454] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7378b40e-977d-4a4a-9489-afebfc4cbd59 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.984424] env[65788]: DEBUG oslo_vmware.api [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1243.984424] env[65788]: value = "task-4663378" [ 1243.984424] env[65788]: _type = "Task" [ 1243.984424] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.992948] env[65788]: DEBUG oslo_vmware.api [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663378, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.170965] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663375, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645867} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.171312] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 67934662-1b06-47b4-9fdf-d8f6ba17b86c/67934662-1b06-47b4-9fdf-d8f6ba17b86c.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1244.171573] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1244.171882] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0e1519e-b386-4e40-a642-25c2de5c6e72 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.181193] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1244.181193] env[65788]: value = "task-4663379" [ 1244.181193] env[65788]: _type = "Task" [ 1244.181193] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.190699] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663379, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.494983] env[65788]: DEBUG oslo_vmware.api [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663378, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178622} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.495267] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1244.495452] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1244.495623] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1244.495804] env[65788]: INFO nova.compute.manager [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: f8c57215-ade6-424b-be8c-075998a764af] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1244.496055] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1244.496258] env[65788]: DEBUG nova.compute.manager [-] [instance: f8c57215-ade6-424b-be8c-075998a764af] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1244.496360] env[65788]: DEBUG nova.network.neutron [-] [instance: f8c57215-ade6-424b-be8c-075998a764af] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1244.496603] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.497145] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1244.497398] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1244.532070] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1244.692038] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663379, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204881} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.692038] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1244.692412] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2540ea21-3ab0-4a74-a36a-f468c758e391 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.715497] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] 67934662-1b06-47b4-9fdf-d8f6ba17b86c/67934662-1b06-47b4-9fdf-d8f6ba17b86c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1244.715903] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27e2a514-0039-488b-9bb1-029acd43a428 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.736779] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1244.736779] env[65788]: value = "task-4663380" [ 1244.736779] env[65788]: _type = "Task" [ 1244.736779] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.745264] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.938119] env[65788]: DEBUG nova.compute.manager [req-91bda9fd-0a7b-459f-a2f9-f813b6d794db req-fd2ea3bd-06b5-49fd-9011-28308f805a49 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Received event network-vif-deleted-fda83fac-56b6-4ab0-824a-7af792b7c8c7 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1244.938119] env[65788]: INFO nova.compute.manager [req-91bda9fd-0a7b-459f-a2f9-f813b6d794db req-fd2ea3bd-06b5-49fd-9011-28308f805a49 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Neutron deleted interface fda83fac-56b6-4ab0-824a-7af792b7c8c7; detaching it from the instance and deleting it from the info cache [ 1244.938119] env[65788]: DEBUG nova.network.neutron [req-91bda9fd-0a7b-459f-a2f9-f813b6d794db req-fd2ea3bd-06b5-49fd-9011-28308f805a49 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1245.247602] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663380, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.418415] env[65788]: DEBUG nova.network.neutron [-] [instance: f8c57215-ade6-424b-be8c-075998a764af] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1245.442093] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3ffce20-3218-41a9-b91c-935bab8fd90b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.455298] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed947f0-8a51-4521-b834-1cba936af37f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.487835] env[65788]: DEBUG nova.compute.manager [req-91bda9fd-0a7b-459f-a2f9-f813b6d794db req-fd2ea3bd-06b5-49fd-9011-28308f805a49 service nova] [instance: f8c57215-ade6-424b-be8c-075998a764af] Detach interface failed, port_id=fda83fac-56b6-4ab0-824a-7af792b7c8c7, reason: Instance f8c57215-ade6-424b-be8c-075998a764af could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1245.750521] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663380, 'name': ReconfigVM_Task, 'duration_secs': 0.876217} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.750818] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Reconfigured VM instance instance-00000079 to attach disk [datastore2] 67934662-1b06-47b4-9fdf-d8f6ba17b86c/67934662-1b06-47b4-9fdf-d8f6ba17b86c.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1245.751552] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d21bbd6a-4d9c-41a1-9ea0-bf9b78f0b977 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.759927] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1245.759927] env[65788]: value = "task-4663381" [ 1245.759927] env[65788]: _type = "Task" [ 1245.759927] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.769291] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663381, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.922494] env[65788]: INFO nova.compute.manager [-] [instance: f8c57215-ade6-424b-be8c-075998a764af] Took 1.43 seconds to deallocate network for instance. [ 1246.270065] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663381, 'name': Rename_Task, 'duration_secs': 0.469785} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.270378] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1246.270505] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a39b7c03-401e-440b-892b-5b5779d87c4c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.278314] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1246.278314] env[65788]: value = "task-4663382" [ 1246.278314] env[65788]: _type = "Task" [ 1246.278314] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.287640] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.429396] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.429686] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.429954] env[65788]: DEBUG nova.objects.instance [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'resources' on Instance uuid f8c57215-ade6-424b-be8c-075998a764af {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1246.789114] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663382, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.053481] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811fdc4c-fd06-4474-9cfd-5971045c35ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.061917] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d41b01-c34a-4da7-af8c-a65a6b941734 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.093359] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca35889-2eba-4b90-b41c-faf5ed475eff {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.101556] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a165e695-b6a9-4bb9-af7a-ef23f2d4187c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.116947] env[65788]: DEBUG nova.compute.provider_tree [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1247.292050] env[65788]: DEBUG oslo_vmware.api [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663382, 'name': PowerOnVM_Task, 'duration_secs': 0.775716} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.292050] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1247.292050] env[65788]: INFO nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Took 8.91 seconds to spawn the instance on the hypervisor. [ 1247.292050] env[65788]: DEBUG nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1247.292958] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932b04f6-bdab-4c4c-8f33-0194b0e35846 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.620665] env[65788]: DEBUG nova.scheduler.client.report [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1247.810560] env[65788]: INFO nova.compute.manager [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Took 13.68 seconds to build instance. [ 1248.126019] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.147230] env[65788]: INFO nova.scheduler.client.report [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted allocations for instance f8c57215-ade6-424b-be8c-075998a764af [ 1248.312848] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a4488b3-d62f-498b-aef9-ac17e9ae8fa4 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.188s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.657851] env[65788]: DEBUG oslo_concurrency.lockutils [None req-400f1708-8955-43bc-9efa-dc5bdc3b7b0e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "f8c57215-ade6-424b-be8c-075998a764af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.788s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.746575] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.746790] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.747032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.747263] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.747436] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1248.749884] env[65788]: INFO nova.compute.manager [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Terminating instance [ 1249.253541] env[65788]: DEBUG nova.compute.manager [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1249.253874] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1249.254910] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71267816-38a6-49ff-9c7f-ef5a57c4dc97 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.262767] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1249.263032] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9deaefcd-8d03-41fe-98af-b2971b352be1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.270281] env[65788]: DEBUG oslo_vmware.api [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1249.270281] env[65788]: value = "task-4663383" [ 1249.270281] env[65788]: _type = "Task" [ 1249.270281] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.279273] env[65788]: DEBUG oslo_vmware.api [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.780901] env[65788]: DEBUG oslo_vmware.api [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663383, 'name': PowerOffVM_Task, 'duration_secs': 0.215814} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.781218] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1249.781342] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1249.781604] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3de338ff-4895-431a-a871-b887f631c840 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.848901] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1249.849179] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1249.849377] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore2] 67934662-1b06-47b4-9fdf-d8f6ba17b86c {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1249.849698] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-426329af-4dbe-4c87-a71d-e7e04ecd8140 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.857204] env[65788]: DEBUG oslo_vmware.api [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1249.857204] env[65788]: value = "task-4663385" [ 1249.857204] env[65788]: _type = "Task" [ 1249.857204] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.866430] env[65788]: DEBUG oslo_vmware.api [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.367338] env[65788]: DEBUG oslo_vmware.api [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127646} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.367598] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1250.367810] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1250.367957] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1250.368150] env[65788]: INFO nova.compute.manager [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1250.368406] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1250.368606] env[65788]: DEBUG nova.compute.manager [-] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1250.368702] env[65788]: DEBUG nova.network.neutron [-] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1250.368959] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.369500] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1250.369757] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1250.405212] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1250.660041] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.660287] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1250.662356] env[65788]: DEBUG nova.compute.manager [req-c8a37590-0349-4a8a-8b45-8a31cda4ea3b req-bef1e042-12bb-43eb-a8a4-cfb5b798adeb service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Received event network-vif-deleted-b9563bd8-34ff-48e6-9804-3643714bb99a {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1250.662541] env[65788]: INFO nova.compute.manager [req-c8a37590-0349-4a8a-8b45-8a31cda4ea3b req-bef1e042-12bb-43eb-a8a4-cfb5b798adeb service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Neutron deleted interface b9563bd8-34ff-48e6-9804-3643714bb99a; detaching it from the instance and deleting it from the info cache [ 1250.662813] env[65788]: DEBUG nova.network.neutron [req-c8a37590-0349-4a8a-8b45-8a31cda4ea3b req-bef1e042-12bb-43eb-a8a4-cfb5b798adeb service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1251.140547] env[65788]: DEBUG nova.network.neutron [-] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1251.165301] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1251.168534] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca66f0b4-6804-4504-9234-b89c7d794f7d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.179159] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5bd7a5-8ff8-43b1-9a55-c789d6671806 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.216774] env[65788]: DEBUG nova.compute.manager [req-c8a37590-0349-4a8a-8b45-8a31cda4ea3b req-bef1e042-12bb-43eb-a8a4-cfb5b798adeb service nova] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Detach interface failed, port_id=b9563bd8-34ff-48e6-9804-3643714bb99a, reason: Instance 67934662-1b06-47b4-9fdf-d8f6ba17b86c could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1251.643067] env[65788]: INFO nova.compute.manager [-] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Took 1.27 seconds to deallocate network for instance. [ 1251.694563] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1251.694904] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1251.696571] env[65788]: INFO nova.compute.claims [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1252.149506] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1252.812587] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e16748-9e37-4281-b1e6-61d05e76a1b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.822664] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352697c4-b9dd-47d1-9e0a-44be3f1324db {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.854870] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7328cba5-c329-4148-8d55-57d62af8450e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.863399] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e0064e-4008-45b6-a069-e284315030f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.877839] env[65788]: DEBUG nova.compute.provider_tree [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1253.381542] env[65788]: DEBUG nova.scheduler.client.report [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1253.480953] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.481237] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.482095] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.482095] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1253.886644] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.192s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.887221] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1253.889981] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.741s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1253.890265] env[65788]: DEBUG nova.objects.instance [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid 67934662-1b06-47b4-9fdf-d8f6ba17b86c {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1254.394234] env[65788]: DEBUG nova.compute.utils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1254.398750] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1254.399101] env[65788]: DEBUG nova.network.neutron [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1254.399282] env[65788]: WARNING neutronclient.v2_0.client [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1254.399757] env[65788]: WARNING neutronclient.v2_0.client [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1254.401206] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1254.401671] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1254.454762] env[65788]: DEBUG nova.policy [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1254.489149] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bcee91b-0947-438c-a76e-68a52b380df8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.497662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269f005f-ec0f-44ed-9c4e-695df21029e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.531048] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb3e595-a76a-4cf1-9d7a-ab7e9b3b9365 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.539673] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4338e553-50bf-4985-a691-11befd41f723 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.554082] env[65788]: DEBUG nova.compute.provider_tree [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.638300] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1254.767604] env[65788]: DEBUG nova.network.neutron [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Successfully created port: 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1254.899869] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1255.057980] env[65788]: DEBUG nova.scheduler.client.report [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1255.564112] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1255.588052] env[65788]: INFO nova.scheduler.client.report [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance 67934662-1b06-47b4-9fdf-d8f6ba17b86c [ 1255.633087] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.637779] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1255.637928] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1255.909665] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1255.935143] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1255.935416] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1255.935572] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1255.935752] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1255.935930] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1255.936105] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1255.936321] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1255.936477] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1255.936640] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1255.936802] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1255.936976] env[65788]: DEBUG nova.virt.hardware [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1255.937886] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b7926f-4f54-467b-968b-d2f1e754cf2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.946720] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de77a1f-d060-4dc3-96c1-555946f028ef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.096676] env[65788]: DEBUG oslo_concurrency.lockutils [None req-12a6045e-ba26-4428-832b-8feab82c5a5d tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "67934662-1b06-47b4-9fdf-d8f6ba17b86c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.350s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1256.154164] env[65788]: DEBUG nova.compute.manager [req-5dce9b00-a917-49b6-9a35-39954cdd50ca req-6d8f15a2-dc89-4d50-83ed-85dc3133a63b service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-vif-plugged-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1256.154485] env[65788]: DEBUG oslo_concurrency.lockutils [req-5dce9b00-a917-49b6-9a35-39954cdd50ca req-6d8f15a2-dc89-4d50-83ed-85dc3133a63b service nova] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1256.154604] env[65788]: DEBUG oslo_concurrency.lockutils [req-5dce9b00-a917-49b6-9a35-39954cdd50ca req-6d8f15a2-dc89-4d50-83ed-85dc3133a63b service nova] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1256.154873] env[65788]: DEBUG oslo_concurrency.lockutils [req-5dce9b00-a917-49b6-9a35-39954cdd50ca req-6d8f15a2-dc89-4d50-83ed-85dc3133a63b service nova] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1256.155194] env[65788]: DEBUG nova.compute.manager [req-5dce9b00-a917-49b6-9a35-39954cdd50ca req-6d8f15a2-dc89-4d50-83ed-85dc3133a63b service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] No waiting events found dispatching network-vif-plugged-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1256.155389] env[65788]: WARNING nova.compute.manager [req-5dce9b00-a917-49b6-9a35-39954cdd50ca req-6d8f15a2-dc89-4d50-83ed-85dc3133a63b service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received unexpected event network-vif-plugged-0c1d654b-57fe-46a4-a1f9-63029c094c97 for instance with vm_state building and task_state spawning. [ 1256.260035] env[65788]: DEBUG nova.network.neutron [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Successfully updated port: 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1256.632347] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1256.762546] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.762769] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1256.762950] env[65788]: DEBUG nova.network.neutron [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1256.879960] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "16dfc544-45f3-486b-bc3d-86836382b07b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1256.880265] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1257.137413] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.267122] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.267538] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.313347] env[65788]: DEBUG nova.network.neutron [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1257.333018] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.333445] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.383011] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1257.398844] env[65788]: WARNING neutronclient.v2_0.client [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1257.399527] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1257.399874] env[65788]: WARNING openstack [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1257.484487] env[65788]: DEBUG nova.network.neutron [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1257.639770] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1257.640140] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1257.640194] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1257.640345] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1257.641265] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb903f1-93d1-411f-a5b9-3280f79feee5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.650196] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68d9bb8-110f-4433-92c5-b53ba3dedb9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.664730] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62683806-53ba-4dd2-bd50-8fe1cddac6a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.671591] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5895c61b-cd75-42e8-8afc-ba24b9d1e9e0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.702152] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179542MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1257.702304] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1257.702524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1257.907041] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1257.987471] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1257.987844] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Instance network_info: |[{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1257.988374] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:4d:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c1d654b-57fe-46a4-a1f9-63029c094c97', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1257.996064] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1257.997098] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1257.997393] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef0d80cc-d9e6-443b-95b5-3a48c675cb51 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.017276] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1258.017276] env[65788]: value = "task-4663386" [ 1258.017276] env[65788]: _type = "Task" [ 1258.017276] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.025475] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663386, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.186886] env[65788]: DEBUG nova.compute.manager [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1258.187120] env[65788]: DEBUG nova.compute.manager [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing instance network info cache due to event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1258.187335] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.187475] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1258.187682] env[65788]: DEBUG nova.network.neutron [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1258.527661] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663386, 'name': CreateVM_Task, 'duration_secs': 0.327153} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.527886] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1258.528382] env[65788]: WARNING neutronclient.v2_0.client [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.528763] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1258.528937] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1258.529266] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1258.529531] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6047546-7259-4d46-9119-f62adcf97bc6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.534875] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1258.534875] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529d6bb0-f5e4-8616-2172-41c585ecfc16" [ 1258.534875] env[65788]: _type = "Task" [ 1258.534875] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.543924] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529d6bb0-f5e4-8616-2172-41c585ecfc16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.690758] env[65788]: WARNING neutronclient.v2_0.client [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.691489] env[65788]: WARNING openstack [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.691840] env[65788]: WARNING openstack [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.730478] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1258.730603] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 27ed2344-c42b-46bc-b51a-20821f67cbf0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1258.730718] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1258.730829] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 222d1456-baad-42ee-bdef-020fb66ebe41 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1258.730938] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance bdd09116-31de-491d-b129-c117f898881e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1258.731056] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4abbb248-12f6-4612-b6ea-01d1ef2104d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1258.793153] env[65788]: WARNING openstack [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.793757] env[65788]: WARNING openstack [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.853731] env[65788]: WARNING neutronclient.v2_0.client [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1258.854431] env[65788]: WARNING openstack [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1258.854783] env[65788]: WARNING openstack [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1258.931724] env[65788]: DEBUG nova.network.neutron [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updated VIF entry in instance network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1258.932106] env[65788]: DEBUG nova.network.neutron [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1259.046556] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529d6bb0-f5e4-8616-2172-41c585ecfc16, 'name': SearchDatastore_Task, 'duration_secs': 0.010572} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.046808] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1259.047052] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1259.047312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.047452] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1259.047687] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1259.048038] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80546387-129f-48fc-b8d2-37638d97eb49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.058423] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1259.058612] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1259.059421] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-548c7e6c-4d63-47ca-91df-bb5edad2d54b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.066053] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1259.066053] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525785a6-9291-73f9-59bb-77dff42b7a01" [ 1259.066053] env[65788]: _type = "Task" [ 1259.066053] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.076037] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525785a6-9291-73f9-59bb-77dff42b7a01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.234230] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 16dfc544-45f3-486b-bc3d-86836382b07b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1259.234472] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1259.234633] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=100GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '6', 'num_vm_active': '5', 'num_task_None': '5', 'num_os_type_None': '6', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '2', 'io_workload': '1', 'num_proj_009f31c7d5bc4d369a8b96e2aa01117a': '1', 'num_proj_94977b75768444138fb914cabddc57c5': '1', 'num_proj_7aa1dec53aa4436daa0f3223b92f9b79': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_proj_676963c25a3741998b10790689242dbb': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1259.332125] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099a5002-6f89-4a18-8543-4512ffd7a49b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.340985] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6a3cc3-525e-4f49-8517-d108f8878852 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.370913] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d22a1cb-84a9-4d38-8d4a-ed245ac9e08e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.378744] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9816482-c28d-46bd-8eba-c8167ecd2229 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.391799] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1259.434849] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1e447f0-7d53-4e54-8505-c7570c6aa7a4 req-0ec2dbc1-4e33-4808-aa59-bea5550e7c39 service nova] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1259.577350] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525785a6-9291-73f9-59bb-77dff42b7a01, 'name': SearchDatastore_Task, 'duration_secs': 0.010937} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.578217] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-196dc37d-45ab-420f-97b4-a84c4269cb57 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.584717] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1259.584717] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c96aa-17b6-e3af-b1c2-59f6134416b3" [ 1259.584717] env[65788]: _type = "Task" [ 1259.584717] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.594283] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c96aa-17b6-e3af-b1c2-59f6134416b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.649322] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1259.649567] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1259.895065] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1260.096202] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522c96aa-17b6-e3af-b1c2-59f6134416b3, 'name': SearchDatastore_Task, 'duration_secs': 0.009869} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.096468] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1260.096727] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4abbb248-12f6-4612-b6ea-01d1ef2104d7/4abbb248-12f6-4612-b6ea-01d1ef2104d7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1260.096988] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b095fc8-d05f-4fe8-9e33-f0858555d104 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.104729] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1260.104729] env[65788]: value = "task-4663387" [ 1260.104729] env[65788]: _type = "Task" [ 1260.104729] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.113303] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.152653] env[65788]: DEBUG nova.compute.utils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1260.400243] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1260.400489] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.698s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1260.400791] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.494s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1260.403094] env[65788]: INFO nova.compute.claims [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1260.615085] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466201} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.615352] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 4abbb248-12f6-4612-b6ea-01d1ef2104d7/4abbb248-12f6-4612-b6ea-01d1ef2104d7.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1260.615577] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1260.615865] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06fa82af-b54a-4331-a1ad-2f88659c6412 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.622302] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1260.622302] env[65788]: value = "task-4663388" [ 1260.622302] env[65788]: _type = "Task" [ 1260.622302] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.631791] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663388, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.655703] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.150276] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663388, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058225} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.150276] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1261.150276] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e150547-e2f2-47ea-bc6c-01d3fc7f7918 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.158055] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Reconfiguring VM instance instance-0000007a to attach disk [datastore2] 4abbb248-12f6-4612-b6ea-01d1ef2104d7/4abbb248-12f6-4612-b6ea-01d1ef2104d7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1261.158373] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59218226-c4d9-4434-9abe-14983a8eb8f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.179455] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1261.179455] env[65788]: value = "task-4663389" [ 1261.179455] env[65788]: _type = "Task" [ 1261.179455] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.189822] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663389, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.478736] env[65788]: INFO nova.compute.manager [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Rebuilding instance [ 1261.522401] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3ea24b-4728-4d85-a595-63607e3edf49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.527925] env[65788]: DEBUG nova.compute.manager [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1261.528761] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118c3c63-3fe6-49b8-b7ef-727ae640d9b1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.534387] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249bae07-aee0-454b-9327-880d0ecd3328 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.570414] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624c6fa8-dff8-4ca0-aebe-01f69a94b9aa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.578202] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68685fc7-4aeb-4a3c-9391-1bf24b03c637 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.593696] env[65788]: DEBUG nova.compute.provider_tree [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1261.691889] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663389, 'name': ReconfigVM_Task, 'duration_secs': 0.283346} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.693025] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Reconfigured VM instance instance-0000007a to attach disk [datastore2] 4abbb248-12f6-4612-b6ea-01d1ef2104d7/4abbb248-12f6-4612-b6ea-01d1ef2104d7.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1261.693232] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e681e7d3-4998-4781-9fd2-bd0f67089cd7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.701235] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1261.701235] env[65788]: value = "task-4663390" [ 1261.701235] env[65788]: _type = "Task" [ 1261.701235] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.711810] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663390, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.751473] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.751803] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.752186] env[65788]: INFO nova.compute.manager [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Attaching volume caf3456e-1c5f-444d-9611-56c217aab77d to /dev/sdb [ 1261.792590] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0eb7cf-42da-483b-a4f5-624d18431a7c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.800634] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93f1d84-a471-4b9d-92f4-06508515603e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.818406] env[65788]: DEBUG nova.virt.block_device [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updating existing volume attachment record: aaebfad5-0201-4f55-be22-668451cf45d6 {{(pid=65788) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1262.097807] env[65788]: DEBUG nova.scheduler.client.report [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1262.215031] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663390, 'name': Rename_Task, 'duration_secs': 0.149673} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.215031] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1262.215031] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9862586e-a795-4456-adfe-41e9dfb55c9f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.223515] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1262.223515] env[65788]: value = "task-4663392" [ 1262.223515] env[65788]: _type = "Task" [ 1262.223515] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.233334] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.547649] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1262.548100] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02a548d8-4abf-40fc-84ef-8ed09c3b984a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.556889] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1262.556889] env[65788]: value = "task-4663393" [ 1262.556889] env[65788]: _type = "Task" [ 1262.556889] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.569384] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.602741] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.202s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1262.603676] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1262.736721] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663392, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.066867] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663393, 'name': PowerOffVM_Task, 'duration_secs': 0.214756} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.067139] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1263.067373] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1263.068151] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea397226-af45-43e1-a721-0fd6df5328c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.075254] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1263.075476] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2b53e8f-e0ae-4720-a0f2-aba1d9351b9e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.110880] env[65788]: DEBUG nova.compute.utils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1263.112459] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1263.112657] env[65788]: DEBUG nova.network.neutron [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1263.113089] env[65788]: WARNING neutronclient.v2_0.client [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.113463] env[65788]: WARNING neutronclient.v2_0.client [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1263.114072] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1263.114427] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1263.142025] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1263.142274] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1263.142501] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleting the datastore file [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1263.142811] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99ced743-4969-4bf2-8919-1c73e996cdb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.149594] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1263.149594] env[65788]: value = "task-4663395" [ 1263.149594] env[65788]: _type = "Task" [ 1263.149594] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.159297] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.162940] env[65788]: DEBUG nova.policy [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6991cd60658e499cbce5da63f5b798e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '009f31c7d5bc4d369a8b96e2aa01117a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1263.237237] env[65788]: DEBUG oslo_vmware.api [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663392, 'name': PowerOnVM_Task, 'duration_secs': 0.750807} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.237574] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1263.237720] env[65788]: INFO nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Took 7.33 seconds to spawn the instance on the hypervisor. [ 1263.237895] env[65788]: DEBUG nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1263.238727] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d333c62-4b43-4147-be11-ddfc8d7ffd82 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.463217] env[65788]: DEBUG nova.network.neutron [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Successfully created port: d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1263.622736] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1263.660497] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138457} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.661449] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1263.661645] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1263.661822] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1263.756808] env[65788]: INFO nova.compute.manager [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Took 12.09 seconds to build instance. [ 1264.258551] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad0cfa5d-4e0b-4e0d-9f10-716ca50fb6a2 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.598s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1264.633273] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1264.660518] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1264.660780] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1264.660946] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1264.661140] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1264.661292] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1264.661438] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1264.661644] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1264.661802] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1264.661965] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1264.662142] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1264.662314] env[65788]: DEBUG nova.virt.hardware [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1264.663424] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f3c14f-39cc-4426-9597-ded792ed6051 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.674750] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe77f2c-dbd0-4503-be11-753ff1f6eeee {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.696360] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1264.696660] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1264.696746] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1264.696935] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1264.697098] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1264.697250] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1264.697456] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1264.697614] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1264.697778] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1264.697939] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1264.698133] env[65788]: DEBUG nova.virt.hardware [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1264.698918] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a651473-3c2a-4445-8512-c13bbf7b0b5a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.707107] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a43d0b8-a445-4eec-a168-726345aa1620 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.721312] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:50:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fff8c844-a1fe-4657-9ab0-99fa3a35d346', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1264.728847] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1264.729494] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1264.729729] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28bb6af3-e6de-4481-9b35-d7e51c884dca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.753300] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1264.753300] env[65788]: value = "task-4663397" [ 1264.753300] env[65788]: _type = "Task" [ 1264.753300] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.762252] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663397, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.862196] env[65788]: DEBUG nova.compute.manager [req-f1a60713-ab47-42c7-baa0-2d74d42f228d req-07eba128-aeef-49ac-b4e9-250ac597fcd4 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Received event network-vif-plugged-d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1264.862196] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1a60713-ab47-42c7-baa0-2d74d42f228d req-07eba128-aeef-49ac-b4e9-250ac597fcd4 service nova] Acquiring lock "16dfc544-45f3-486b-bc3d-86836382b07b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1264.862196] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1a60713-ab47-42c7-baa0-2d74d42f228d req-07eba128-aeef-49ac-b4e9-250ac597fcd4 service nova] Lock "16dfc544-45f3-486b-bc3d-86836382b07b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1264.862196] env[65788]: DEBUG oslo_concurrency.lockutils [req-f1a60713-ab47-42c7-baa0-2d74d42f228d req-07eba128-aeef-49ac-b4e9-250ac597fcd4 service nova] Lock "16dfc544-45f3-486b-bc3d-86836382b07b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1264.862449] env[65788]: DEBUG nova.compute.manager [req-f1a60713-ab47-42c7-baa0-2d74d42f228d req-07eba128-aeef-49ac-b4e9-250ac597fcd4 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] No waiting events found dispatching network-vif-plugged-d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1264.862734] env[65788]: WARNING nova.compute.manager [req-f1a60713-ab47-42c7-baa0-2d74d42f228d req-07eba128-aeef-49ac-b4e9-250ac597fcd4 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Received unexpected event network-vif-plugged-d7eca264-2c75-44a4-a968-213e84c89d5b for instance with vm_state building and task_state spawning. [ 1264.948428] env[65788]: DEBUG nova.network.neutron [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Successfully updated port: d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1265.020572] env[65788]: DEBUG nova.compute.manager [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1265.020745] env[65788]: DEBUG nova.compute.manager [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing instance network info cache due to event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1265.021581] env[65788]: DEBUG oslo_concurrency.lockutils [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.021581] env[65788]: DEBUG oslo_concurrency.lockutils [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1265.021581] env[65788]: DEBUG nova.network.neutron [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1265.267032] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663397, 'name': CreateVM_Task, 'duration_secs': 0.356306} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.267032] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1265.267032] env[65788]: WARNING neutronclient.v2_0.client [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release.: NotImplementedError [ 1265.267032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.267032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1265.267032] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1265.267032] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69c1aded-751f-435f-9616-5a201c687189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.272437] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1265.272437] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cd6f2a-6bdc-f15d-664a-2649da4f25cf" [ 1265.272437] env[65788]: _type = "Task" [ 1265.272437] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.282858] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cd6f2a-6bdc-f15d-664a-2649da4f25cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.451483] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "refresh_cache-16dfc544-45f3-486b-bc3d-86836382b07b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.451701] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "refresh_cache-16dfc544-45f3-486b-bc3d-86836382b07b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1265.451883] env[65788]: DEBUG nova.network.neutron [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1265.523887] env[65788]: WARNING neutronclient.v2_0.client [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1265.524628] env[65788]: WARNING openstack [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.525023] env[65788]: WARNING openstack [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.667309] env[65788]: WARNING openstack [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.667702] env[65788]: WARNING openstack [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.730572] env[65788]: WARNING neutronclient.v2_0.client [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1265.731246] env[65788]: WARNING openstack [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.731590] env[65788]: WARNING openstack [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1265.788322] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cd6f2a-6bdc-f15d-664a-2649da4f25cf, 'name': SearchDatastore_Task, 'duration_secs': 0.01103} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.788642] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.788849] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1265.789104] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.789646] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1265.789646] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1265.789799] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07577071-8152-4ae3-8ab4-ed5c366f5b9d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.799249] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1265.799406] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1265.800136] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f867a2d-f206-488a-97ff-2d112e62bbfd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.806301] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1265.806301] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a54ced-9fe5-486f-1035-3905195d8548" [ 1265.806301] env[65788]: _type = "Task" [ 1265.806301] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.814459] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a54ced-9fe5-486f-1035-3905195d8548, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.816626] env[65788]: DEBUG nova.network.neutron [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updated VIF entry in instance network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1265.817038] env[65788]: DEBUG nova.network.neutron [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1265.955166] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1265.955602] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.000763] env[65788]: DEBUG nova.network.neutron [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1266.020229] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1266.020700] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.032112] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1266.032325] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1266.083746] env[65788]: WARNING neutronclient.v2_0.client [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1266.084432] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1266.084786] env[65788]: WARNING openstack [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1266.165928] env[65788]: DEBUG nova.network.neutron [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Updating instance_info_cache with network_info: [{"id": "d7eca264-2c75-44a4-a968-213e84c89d5b", "address": "fa:16:3e:8d:7a:fa", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7eca264-2c", "ovs_interfaceid": "d7eca264-2c75-44a4-a968-213e84c89d5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1266.319220] env[65788]: DEBUG oslo_concurrency.lockutils [req-bfdf9546-69f3-42d2-a9c7-557f0ccc8388 req-14384655-2709-4ddc-9554-5af2595eadfe service nova] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1266.319599] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52a54ced-9fe5-486f-1035-3905195d8548, 'name': SearchDatastore_Task, 'duration_secs': 0.009398} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.320315] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-055b2a56-3d0e-4249-938c-0a37cd145b7b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.325834] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1266.325834] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52568bca-f06a-829b-fb5d-5be17d70777e" [ 1266.325834] env[65788]: _type = "Task" [ 1266.325834] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.333993] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52568bca-f06a-829b-fb5d-5be17d70777e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.365699] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Volume attach. Driver type: vmdk {{(pid=65788) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1266.366088] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910450', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'name': 'volume-caf3456e-1c5f-444d-9611-56c217aab77d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'serial': 'caf3456e-1c5f-444d-9611-56c217aab77d'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1266.367075] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508f7f5b-995e-4c22-a923-47abcbbf728a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.385667] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0a425d-c356-4db2-909b-17a04bfe2eac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.411120] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] volume-caf3456e-1c5f-444d-9611-56c217aab77d/volume-caf3456e-1c5f-444d-9611-56c217aab77d.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1266.411453] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b5e3a2b-8426-41e3-860f-4c9563958f99 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.430105] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1266.430105] env[65788]: value = "task-4663398" [ 1266.430105] env[65788]: _type = "Task" [ 1266.430105] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.439213] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663398, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.534985] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1266.668394] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "refresh_cache-16dfc544-45f3-486b-bc3d-86836382b07b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1266.668774] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Instance network_info: |[{"id": "d7eca264-2c75-44a4-a968-213e84c89d5b", "address": "fa:16:3e:8d:7a:fa", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7eca264-2c", "ovs_interfaceid": "d7eca264-2c75-44a4-a968-213e84c89d5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1266.669296] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:7a:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca99f7a1-6365-4d3c-af16-1b1c1288091e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd7eca264-2c75-44a4-a968-213e84c89d5b', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1266.676812] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1266.676975] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1266.677245] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4a5a1e1-145b-4e8c-a37a-72f307177c4e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.698585] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1266.698585] env[65788]: value = "task-4663399" [ 1266.698585] env[65788]: _type = "Task" [ 1266.698585] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.707695] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663399, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.838443] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52568bca-f06a-829b-fb5d-5be17d70777e, 'name': SearchDatastore_Task, 'duration_secs': 0.01029} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.838923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1266.839171] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1266.839504] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32c0f4d4-e487-459b-8943-ff14bf057ea9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.851347] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1266.851347] env[65788]: value = "task-4663400" [ 1266.851347] env[65788]: _type = "Task" [ 1266.851347] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.860553] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663400, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.896644] env[65788]: DEBUG nova.compute.manager [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Received event network-changed-d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1266.897289] env[65788]: DEBUG nova.compute.manager [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Refreshing instance network info cache due to event network-changed-d7eca264-2c75-44a4-a968-213e84c89d5b. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1266.897289] env[65788]: DEBUG oslo_concurrency.lockutils [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Acquiring lock "refresh_cache-16dfc544-45f3-486b-bc3d-86836382b07b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1266.897289] env[65788]: DEBUG oslo_concurrency.lockutils [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Acquired lock "refresh_cache-16dfc544-45f3-486b-bc3d-86836382b07b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1266.897479] env[65788]: DEBUG nova.network.neutron [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Refreshing network info cache for port d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1266.943588] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.059976] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1267.060470] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1267.062335] env[65788]: INFO nova.compute.claims [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1267.210466] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663399, 'name': CreateVM_Task, 'duration_secs': 0.327383} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.210680] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1267.211297] env[65788]: WARNING neutronclient.v2_0.client [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.211635] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1267.211787] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1267.212124] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1267.212404] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74eb64df-b3cb-4569-bfcc-8b538052991e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.218373] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1267.218373] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cf3bfa-2ae9-8df1-c1ad-0629f2e72d60" [ 1267.218373] env[65788]: _type = "Task" [ 1267.218373] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.227727] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cf3bfa-2ae9-8df1-c1ad-0629f2e72d60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.364975] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663400, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471858} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.365439] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1267.365749] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1267.366160] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76e1a89e-9ff8-4041-bc90-6913751472f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.375308] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1267.375308] env[65788]: value = "task-4663401" [ 1267.375308] env[65788]: _type = "Task" [ 1267.375308] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.389510] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663401, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.400567] env[65788]: WARNING neutronclient.v2_0.client [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.401596] env[65788]: WARNING openstack [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.402179] env[65788]: WARNING openstack [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.445497] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663398, 'name': ReconfigVM_Task, 'duration_secs': 0.626696} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.448775] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Reconfigured VM instance instance-00000077 to attach disk [datastore2] volume-caf3456e-1c5f-444d-9611-56c217aab77d/volume-caf3456e-1c5f-444d-9611-56c217aab77d.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1267.456720] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84f864a9-a844-40e3-bfe1-d1ccf45de550 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.480515] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1267.480515] env[65788]: value = "task-4663402" [ 1267.480515] env[65788]: _type = "Task" [ 1267.480515] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.492566] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663402, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.518116] env[65788]: WARNING openstack [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.518717] env[65788]: WARNING openstack [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.583305] env[65788]: WARNING neutronclient.v2_0.client [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1267.584045] env[65788]: WARNING openstack [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1267.584425] env[65788]: WARNING openstack [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1267.666545] env[65788]: DEBUG nova.network.neutron [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Updated VIF entry in instance network info cache for port d7eca264-2c75-44a4-a968-213e84c89d5b. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1267.666907] env[65788]: DEBUG nova.network.neutron [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Updating instance_info_cache with network_info: [{"id": "d7eca264-2c75-44a4-a968-213e84c89d5b", "address": "fa:16:3e:8d:7a:fa", "network": {"id": "cf36051a-49c3-49ad-a6c9-a4146537d4e2", "bridge": "br-int", "label": "tempest-ServersTestJSON-345655545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "009f31c7d5bc4d369a8b96e2aa01117a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca99f7a1-6365-4d3c-af16-1b1c1288091e", "external-id": "cl2-zone-334", "segmentation_id": 334, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd7eca264-2c", "ovs_interfaceid": "d7eca264-2c75-44a4-a968-213e84c89d5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1267.730232] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52cf3bfa-2ae9-8df1-c1ad-0629f2e72d60, 'name': SearchDatastore_Task, 'duration_secs': 0.054511} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.730555] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1267.730789] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1267.731036] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1267.731191] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1267.731372] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1267.731646] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aeed0403-a04c-44df-a591-8f1af41fb16b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.741915] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1267.742204] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1267.743118] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99429404-c4ee-459c-96ef-b0b80789d8f3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.749787] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1267.749787] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b049f-1750-9a1f-0736-66c05193d252" [ 1267.749787] env[65788]: _type = "Task" [ 1267.749787] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.758939] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b049f-1750-9a1f-0736-66c05193d252, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.885587] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663401, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06875} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.885854] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1267.886807] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8a15ac-2ad2-4b52-a2e2-64a063d4b64f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.909615] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1267.909939] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8568c0a7-44e6-41db-9f9f-e5005e3bb6d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.933778] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1267.933778] env[65788]: value = "task-4663403" [ 1267.933778] env[65788]: _type = "Task" [ 1267.933778] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.942866] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663403, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.991037] env[65788]: DEBUG oslo_vmware.api [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663402, 'name': ReconfigVM_Task, 'duration_secs': 0.307117} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.991414] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910450', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'name': 'volume-caf3456e-1c5f-444d-9611-56c217aab77d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'serial': 'caf3456e-1c5f-444d-9611-56c217aab77d'} {{(pid=65788) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1268.169211] env[65788]: DEBUG oslo_concurrency.lockutils [req-d1d44df3-1515-4833-9df6-63e7af942438 req-088f0c4a-3fab-4c6e-b3a6-fff8a08b212a service nova] Releasing lock "refresh_cache-16dfc544-45f3-486b-bc3d-86836382b07b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1268.189045] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c763c190-99ca-4487-ae39-c4c024fe7729 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.197527] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d3d708-6f15-4410-81cb-2e8a90624bae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.231205] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a2caf7-c202-43ad-bcdd-a7f1deb8431d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.239241] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60de3ed5-781f-4b23-8290-f17568f1d5fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.252891] env[65788]: DEBUG nova.compute.provider_tree [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.262884] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529b049f-1750-9a1f-0736-66c05193d252, 'name': SearchDatastore_Task, 'duration_secs': 0.011209} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.264232] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5dd0e3e-5dd4-4725-aee0-d672264a9bc9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.271203] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1268.271203] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eec85f-69bb-e6ee-a056-b2f5b45be7e3" [ 1268.271203] env[65788]: _type = "Task" [ 1268.271203] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.279693] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eec85f-69bb-e6ee-a056-b2f5b45be7e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.446068] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663403, 'name': ReconfigVM_Task, 'duration_secs': 0.274312} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.446446] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Reconfigured VM instance instance-00000078 to attach disk [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41/222d1456-baad-42ee-bdef-020fb66ebe41.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1268.446955] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7e046c8-bd40-4177-ac61-e12583d201ae {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.454566] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1268.454566] env[65788]: value = "task-4663404" [ 1268.454566] env[65788]: _type = "Task" [ 1268.454566] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.463452] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663404, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.758195] env[65788]: DEBUG nova.scheduler.client.report [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1268.783558] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52eec85f-69bb-e6ee-a056-b2f5b45be7e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010669} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.784020] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1268.784410] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 16dfc544-45f3-486b-bc3d-86836382b07b/16dfc544-45f3-486b-bc3d-86836382b07b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1268.784536] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38bee57e-8ba9-4d07-8928-fab27ad4f4d7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.793583] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1268.793583] env[65788]: value = "task-4663405" [ 1268.793583] env[65788]: _type = "Task" [ 1268.793583] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.804093] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663405, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.966670] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663404, 'name': Rename_Task, 'duration_secs': 0.156662} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.967011] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1268.967301] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-976716ad-235d-4856-a270-5f75831f95b4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.974470] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1268.974470] env[65788]: value = "task-4663406" [ 1268.974470] env[65788]: _type = "Task" [ 1268.974470] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.984150] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.037147] env[65788]: DEBUG nova.objects.instance [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'flavor' on Instance uuid ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1269.247167] env[65788]: DEBUG oslo_concurrency.lockutils [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.265077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.265697] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1269.309614] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663405, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.487432] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663406, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.542682] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e0d31832-085c-40a2-bcff-68466b05e14e tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.791s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.543665] env[65788]: DEBUG oslo_concurrency.lockutils [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.297s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.771700] env[65788]: DEBUG nova.compute.utils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1269.773167] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1269.773361] env[65788]: DEBUG nova.network.neutron [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1269.773679] env[65788]: WARNING neutronclient.v2_0.client [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1269.773971] env[65788]: WARNING neutronclient.v2_0.client [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1269.774560] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1269.774931] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1269.808690] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663405, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553124} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.809268] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 16dfc544-45f3-486b-bc3d-86836382b07b/16dfc544-45f3-486b-bc3d-86836382b07b.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1269.809634] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1269.810029] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4d08b15-1d45-4f35-a8c2-68625ae31d8a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.819081] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1269.819081] env[65788]: value = "task-4663407" [ 1269.819081] env[65788]: _type = "Task" [ 1269.819081] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.829969] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663407, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.837315] env[65788]: DEBUG nova.policy [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1269.986910] env[65788]: DEBUG oslo_vmware.api [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663406, 'name': PowerOnVM_Task, 'duration_secs': 0.523488} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.987316] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1269.987548] env[65788]: DEBUG nova.compute.manager [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1269.988387] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58c8d88-75ce-41eb-94c1-1339481eb567 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.046527] env[65788]: INFO nova.compute.manager [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Detaching volume caf3456e-1c5f-444d-9611-56c217aab77d [ 1270.087361] env[65788]: INFO nova.virt.block_device [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Attempting to driver detach volume caf3456e-1c5f-444d-9611-56c217aab77d from mountpoint /dev/sdb [ 1270.087607] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Volume detach. Driver type: vmdk {{(pid=65788) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1270.087876] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910450', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'name': 'volume-caf3456e-1c5f-444d-9611-56c217aab77d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'serial': 'caf3456e-1c5f-444d-9611-56c217aab77d'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1270.088791] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a583fab1-87c3-49ec-b69e-9c75fb14a2c6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.128292] env[65788]: DEBUG nova.network.neutron [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Successfully created port: b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1270.132539] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812507f8-f03d-4d11-88da-9e4a99269cab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.145075] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30de6d8f-8fc5-4278-8a7c-97f8fcca35bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.179744] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d61b1b8-eb2e-4c5c-87c2-d079eeeae87c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.204175] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] The volume has not been displaced from its original location: [datastore2] volume-caf3456e-1c5f-444d-9611-56c217aab77d/volume-caf3456e-1c5f-444d-9611-56c217aab77d.vmdk. No consolidation needed. {{(pid=65788) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1270.209926] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Reconfiguring VM instance instance-00000077 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1270.210362] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c820f450-9a6b-4c20-9b3c-7e451b2467c1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.230907] env[65788]: DEBUG oslo_vmware.api [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1270.230907] env[65788]: value = "task-4663408" [ 1270.230907] env[65788]: _type = "Task" [ 1270.230907] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.240703] env[65788]: DEBUG oslo_vmware.api [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663408, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.282868] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1270.330476] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663407, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213144} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.331531] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1270.332441] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce766485-47c2-4333-b269-3fe814d6c903 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.359708] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Reconfiguring VM instance instance-0000007b to attach disk [datastore2] 16dfc544-45f3-486b-bc3d-86836382b07b/16dfc544-45f3-486b-bc3d-86836382b07b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1270.360384] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-004391a8-160e-4ca9-894a-741d43d32626 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.384401] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1270.384401] env[65788]: value = "task-4663409" [ 1270.384401] env[65788]: _type = "Task" [ 1270.384401] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.395529] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663409, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.508087] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1270.508401] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1270.508777] env[65788]: DEBUG nova.objects.instance [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1270.742568] env[65788]: DEBUG oslo_vmware.api [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663408, 'name': ReconfigVM_Task, 'duration_secs': 0.322991} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.742871] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Reconfigured VM instance instance-00000077 to detach disk 2001 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1270.747630] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df1848c1-b4f9-4af7-a241-2d739f93240f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.764821] env[65788]: DEBUG oslo_vmware.api [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1270.764821] env[65788]: value = "task-4663410" [ 1270.764821] env[65788]: _type = "Task" [ 1270.764821] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.774234] env[65788]: DEBUG oslo_vmware.api [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663410, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.896369] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.274706] env[65788]: DEBUG oslo_vmware.api [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663410, 'name': ReconfigVM_Task, 'duration_secs': 0.156866} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.275095] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-910450', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'name': 'volume-caf3456e-1c5f-444d-9611-56c217aab77d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0', 'attached_at': '', 'detached_at': '', 'volume_id': 'caf3456e-1c5f-444d-9611-56c217aab77d', 'serial': 'caf3456e-1c5f-444d-9611-56c217aab77d'} {{(pid=65788) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1271.293524] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1271.320152] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1271.320435] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1271.320590] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1271.320769] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1271.320915] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1271.321081] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1271.321318] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1271.321487] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1271.321657] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1271.321826] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1271.322010] env[65788]: DEBUG nova.virt.hardware [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1271.322923] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42a52fc-8be9-4b22-8576-f87dddf87caa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.332079] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4feadf2f-cf0b-4b97-9749-70a70edf9b5d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.395482] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663409, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.517386] env[65788]: DEBUG oslo_concurrency.lockutils [None req-430ec20f-781f-4f8c-ad10-a142ed88756d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1271.613313] env[65788]: DEBUG nova.compute.manager [req-4c56c853-ca80-480a-af44-38f8159c1df3 req-c7e9d575-c4e8-4c9c-8712-ade2794e0a4c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-vif-plugged-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1271.613656] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c56c853-ca80-480a-af44-38f8159c1df3 req-c7e9d575-c4e8-4c9c-8712-ade2794e0a4c service nova] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1271.613905] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c56c853-ca80-480a-af44-38f8159c1df3 req-c7e9d575-c4e8-4c9c-8712-ade2794e0a4c service nova] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1271.614111] env[65788]: DEBUG oslo_concurrency.lockutils [req-4c56c853-ca80-480a-af44-38f8159c1df3 req-c7e9d575-c4e8-4c9c-8712-ade2794e0a4c service nova] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1271.614306] env[65788]: DEBUG nova.compute.manager [req-4c56c853-ca80-480a-af44-38f8159c1df3 req-c7e9d575-c4e8-4c9c-8712-ade2794e0a4c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] No waiting events found dispatching network-vif-plugged-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1271.614478] env[65788]: WARNING nova.compute.manager [req-4c56c853-ca80-480a-af44-38f8159c1df3 req-c7e9d575-c4e8-4c9c-8712-ade2794e0a4c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received unexpected event network-vif-plugged-b6e00388-271a-41b9-bb29-4463d4a692ea for instance with vm_state building and task_state spawning. [ 1271.711896] env[65788]: DEBUG nova.network.neutron [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Successfully updated port: b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1271.824835] env[65788]: DEBUG nova.objects.instance [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'flavor' on Instance uuid ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1271.897430] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663409, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.214866] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1272.215137] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1272.215234] env[65788]: DEBUG nova.network.neutron [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1272.398727] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663409, 'name': ReconfigVM_Task, 'duration_secs': 1.527315} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.399040] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Reconfigured VM instance instance-0000007b to attach disk [datastore2] 16dfc544-45f3-486b-bc3d-86836382b07b/16dfc544-45f3-486b-bc3d-86836382b07b.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1272.399735] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61472d0a-fc0f-41c8-b15d-7098ac04f0d8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.408704] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1272.408704] env[65788]: value = "task-4663411" [ 1272.408704] env[65788]: _type = "Task" [ 1272.408704] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.418639] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663411, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.719341] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1272.720075] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1272.777772] env[65788]: DEBUG nova.network.neutron [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1272.799661] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1272.800084] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1272.832597] env[65788]: DEBUG oslo_concurrency.lockutils [None req-290d2662-aa5a-4b38-b1c9-f79a0ffb11aa tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.289s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1272.878890] env[65788]: WARNING neutronclient.v2_0.client [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1272.879623] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1272.880007] env[65788]: WARNING openstack [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1272.920219] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663411, 'name': Rename_Task, 'duration_secs': 0.157782} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.920549] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1272.920788] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-720463ad-532d-43e1-8ee3-c6302adf3ce1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1272.932922] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1272.932922] env[65788]: value = "task-4663412" [ 1272.932922] env[65788]: _type = "Task" [ 1272.932922] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1272.942941] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663412, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.984422] env[65788]: DEBUG nova.network.neutron [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1273.447303] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663412, 'name': PowerOnVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.487482] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1273.487900] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Instance network_info: |[{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1273.488477] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:9f:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6e00388-271a-41b9-bb29-4463d4a692ea', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1273.500078] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1273.500473] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1273.500857] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5320bb9-5120-4c72-a15a-4aeccf2ab2cb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.529230] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1273.529230] env[65788]: value = "task-4663413" [ 1273.529230] env[65788]: _type = "Task" [ 1273.529230] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.539378] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663413, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.560191] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1273.560481] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1273.560905] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1273.561143] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1273.561340] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1273.563623] env[65788]: INFO nova.compute.manager [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Terminating instance [ 1273.654598] env[65788]: DEBUG nova.compute.manager [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1273.654811] env[65788]: DEBUG nova.compute.manager [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing instance network info cache due to event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1273.655131] env[65788]: DEBUG oslo_concurrency.lockutils [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.655316] env[65788]: DEBUG oslo_concurrency.lockutils [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1273.655590] env[65788]: DEBUG nova.network.neutron [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1273.946467] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663412, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.040986] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663413, 'name': CreateVM_Task, 'duration_secs': 0.502088} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.041290] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1274.041763] env[65788]: WARNING neutronclient.v2_0.client [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.042232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.042408] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1274.042799] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1274.043179] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cd48cc8-b163-499b-8257-6c1ff5c12fa2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.050089] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1274.050089] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527ca2ed-8d09-287a-63a5-de2332dcee29" [ 1274.050089] env[65788]: _type = "Task" [ 1274.050089] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.060497] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527ca2ed-8d09-287a-63a5-de2332dcee29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.067416] env[65788]: DEBUG nova.compute.manager [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1274.067655] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1274.068646] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864f27c-4081-4657-96d5-25d37849b4f8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.077998] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1274.078331] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b914c51a-2997-4272-ae18-a227f76871e3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.088066] env[65788]: DEBUG oslo_vmware.api [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1274.088066] env[65788]: value = "task-4663414" [ 1274.088066] env[65788]: _type = "Task" [ 1274.088066] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.100319] env[65788]: DEBUG oslo_vmware.api [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663414, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.158392] env[65788]: WARNING neutronclient.v2_0.client [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.159110] env[65788]: WARNING openstack [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.159476] env[65788]: WARNING openstack [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.261619] env[65788]: WARNING openstack [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.262057] env[65788]: WARNING openstack [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.325175] env[65788]: WARNING neutronclient.v2_0.client [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1274.325849] env[65788]: WARNING openstack [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1274.326278] env[65788]: WARNING openstack [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1274.413401] env[65788]: DEBUG nova.network.neutron [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updated VIF entry in instance network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1274.413833] env[65788]: DEBUG nova.network.neutron [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1274.444465] env[65788]: DEBUG oslo_vmware.api [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663412, 'name': PowerOnVM_Task, 'duration_secs': 1.045585} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.444733] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1274.444986] env[65788]: INFO nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Took 9.81 seconds to spawn the instance on the hypervisor. [ 1274.445134] env[65788]: DEBUG nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1274.445923] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e50cf12-264c-4860-8472-608dfe9bdeb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.561432] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]527ca2ed-8d09-287a-63a5-de2332dcee29, 'name': SearchDatastore_Task, 'duration_secs': 0.03044} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.561685] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1274.562092] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1274.562371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.562521] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1274.562705] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1274.562993] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b6ea159-602d-408c-befb-8449b28a5487 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.572514] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1274.572775] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1274.573535] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b72d310-2955-4d9d-a350-92b526183511 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.579184] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1274.579184] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d240c8-cb1b-6730-1236-19c8cc1444e7" [ 1274.579184] env[65788]: _type = "Task" [ 1274.579184] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.587374] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d240c8-cb1b-6730-1236-19c8cc1444e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.596481] env[65788]: DEBUG oslo_vmware.api [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663414, 'name': PowerOffVM_Task, 'duration_secs': 0.34143} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.596724] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1274.596887] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1274.597155] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6dac482-bce7-420b-98b9-ae40521bcfc5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.668325] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1274.668581] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1274.668762] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleting the datastore file [datastore2] ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1274.669051] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c56eb5b-2288-46a1-bd50-8456cb109c2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.678193] env[65788]: DEBUG oslo_vmware.api [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for the task: (returnval){ [ 1274.678193] env[65788]: value = "task-4663416" [ 1274.678193] env[65788]: _type = "Task" [ 1274.678193] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.686882] env[65788]: DEBUG oslo_vmware.api [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663416, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.917305] env[65788]: DEBUG oslo_concurrency.lockutils [req-a86a4274-98b4-4e3e-ae60-7ddd6369f70e req-3f4db3bb-c19c-4401-ab35-86f9274bb91c service nova] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1274.962281] env[65788]: INFO nova.compute.manager [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Took 17.07 seconds to build instance. [ 1275.089701] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d240c8-cb1b-6730-1236-19c8cc1444e7, 'name': SearchDatastore_Task, 'duration_secs': 0.009002} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.090550] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e68aa4f7-ee55-43d7-be06-e21c389adc49 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.097472] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1275.097472] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f4a31-a42b-b8cf-3cb9-cb1553a9f8ef" [ 1275.097472] env[65788]: _type = "Task" [ 1275.097472] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.105772] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f4a31-a42b-b8cf-3cb9-cb1553a9f8ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.189903] env[65788]: DEBUG oslo_vmware.api [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Task: {'id': task-4663416, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136227} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.190098] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1275.190209] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1275.190384] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1275.190565] env[65788]: INFO nova.compute.manager [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1275.190834] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1275.191072] env[65788]: DEBUG nova.compute.manager [-] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1275.191154] env[65788]: DEBUG nova.network.neutron [-] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1275.191678] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.191958] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1275.192266] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1275.266661] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1275.466341] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0f3411af-c49e-4dcb-a9e6-7668e665b175 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.585s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1275.611083] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]521f4a31-a42b-b8cf-3cb9-cb1553a9f8ef, 'name': SearchDatastore_Task, 'duration_secs': 0.010226} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.611711] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1275.611983] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] cc9be8c2-a6a8-4b33-9b15-e9ee58e43304/cc9be8c2-a6a8-4b33-9b15-e9ee58e43304.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1275.612544] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-822f5f73-19e9-493d-9498-1ffd3383d6c3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.621840] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1275.621840] env[65788]: value = "task-4663417" [ 1275.621840] env[65788]: _type = "Task" [ 1275.621840] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.632275] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.753195] env[65788]: DEBUG nova.compute.manager [req-f32dc573-5d89-4957-be3b-44539c383072 req-d54a4cbd-67fd-40ee-ba58-231c317126ce service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Received event network-vif-deleted-f42a00fc-23eb-4df0-a834-72750f41e45f {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1275.753195] env[65788]: INFO nova.compute.manager [req-f32dc573-5d89-4957-be3b-44539c383072 req-d54a4cbd-67fd-40ee-ba58-231c317126ce service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Neutron deleted interface f42a00fc-23eb-4df0-a834-72750f41e45f; detaching it from the instance and deleting it from the info cache [ 1275.753195] env[65788]: DEBUG nova.network.neutron [req-f32dc573-5d89-4957-be3b-44539c383072 req-d54a4cbd-67fd-40ee-ba58-231c317126ce service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1276.047375] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "16dfc544-45f3-486b-bc3d-86836382b07b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1276.047776] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1276.047999] env[65788]: DEBUG nova.compute.manager [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1276.049058] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125cf996-9305-493f-8ff1-300fb399f635 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.057164] env[65788]: DEBUG nova.compute.manager [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1276.057796] env[65788]: DEBUG nova.objects.instance [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'flavor' on Instance uuid 16dfc544-45f3-486b-bc3d-86836382b07b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1276.095661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1276.095935] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1276.096354] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "bdd09116-31de-491d-b129-c117f898881e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1276.096612] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1276.096811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1276.099254] env[65788]: INFO nova.compute.manager [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Terminating instance [ 1276.136436] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663417, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.227278] env[65788]: DEBUG nova.network.neutron [-] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1276.256997] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e76e6c1-0052-4698-95d8-0de7e1607652 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.268010] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadab987-3a84-4b22-8a98-5f62b702fcaf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.300065] env[65788]: DEBUG nova.compute.manager [req-f32dc573-5d89-4957-be3b-44539c383072 req-d54a4cbd-67fd-40ee-ba58-231c317126ce service nova] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Detach interface failed, port_id=f42a00fc-23eb-4df0-a834-72750f41e45f, reason: Instance ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1276.602855] env[65788]: DEBUG nova.compute.manager [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1276.603188] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1276.604164] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce26159-ca08-4578-9a39-9abca4f316c1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.613980] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1276.614357] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b85e6769-f5eb-4824-b294-7dec7fd995da {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.621777] env[65788]: DEBUG oslo_vmware.api [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1276.621777] env[65788]: value = "task-4663418" [ 1276.621777] env[65788]: _type = "Task" [ 1276.621777] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.633604] env[65788]: DEBUG oslo_vmware.api [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663418, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.637167] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555398} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.637450] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] cc9be8c2-a6a8-4b33-9b15-e9ee58e43304/cc9be8c2-a6a8-4b33-9b15-e9ee58e43304.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1276.637662] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1276.637928] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cae211f-8224-44b8-837a-32e83ebc7084 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.645865] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1276.645865] env[65788]: value = "task-4663419" [ 1276.645865] env[65788]: _type = "Task" [ 1276.645865] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.655789] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663419, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.731838] env[65788]: INFO nova.compute.manager [-] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Took 1.54 seconds to deallocate network for instance. [ 1277.068636] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1277.069033] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa03b943-5173-4ae3-be77-c24edf58532c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.077344] env[65788]: DEBUG oslo_vmware.api [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1277.077344] env[65788]: value = "task-4663420" [ 1277.077344] env[65788]: _type = "Task" [ 1277.077344] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.087041] env[65788]: DEBUG oslo_vmware.api [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.132300] env[65788]: DEBUG oslo_vmware.api [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663418, 'name': PowerOffVM_Task, 'duration_secs': 0.223083} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.132534] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1277.132705] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1277.133194] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9135fe2c-e7ac-4f50-a578-ec111e46634c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.155935] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663419, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070129} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.156318] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1277.157156] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1f4e87-ecb7-48ca-88cf-900350a86c76 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.184785] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] cc9be8c2-a6a8-4b33-9b15-e9ee58e43304/cc9be8c2-a6a8-4b33-9b15-e9ee58e43304.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1277.185263] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81dfd272-80f4-4101-bc6a-fdfdf010aad1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.210604] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1277.210604] env[65788]: value = "task-4663422" [ 1277.210604] env[65788]: _type = "Task" [ 1277.210604] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.215409] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1277.215649] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1277.215829] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleting the datastore file [datastore1] bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1277.216660] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41492978-fc38-4167-8988-204e9a3661ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.223994] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.226223] env[65788]: DEBUG oslo_vmware.api [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for the task: (returnval){ [ 1277.226223] env[65788]: value = "task-4663423" [ 1277.226223] env[65788]: _type = "Task" [ 1277.226223] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1277.235409] env[65788]: DEBUG oslo_vmware.api [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.241819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1277.242184] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1277.242368] env[65788]: DEBUG nova.objects.instance [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lazy-loading 'resources' on Instance uuid ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1277.589063] env[65788]: DEBUG oslo_vmware.api [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663420, 'name': PowerOffVM_Task, 'duration_secs': 0.294324} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.589381] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1277.589575] env[65788]: DEBUG nova.compute.manager [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1277.590551] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9eb1b3b-3215-4e86-8725-3bac52a81e8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.723714] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663422, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.736439] env[65788]: DEBUG oslo_vmware.api [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Task: {'id': task-4663423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153943} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1277.736748] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1277.736942] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1277.737164] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1277.737378] env[65788]: INFO nova.compute.manager [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] [instance: bdd09116-31de-491d-b129-c117f898881e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1277.737673] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1277.737877] env[65788]: DEBUG nova.compute.manager [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1277.737979] env[65788]: DEBUG nova.network.neutron [-] [instance: bdd09116-31de-491d-b129-c117f898881e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1277.738268] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.738808] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1277.739079] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1277.779345] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1277.873765] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c75bd65-deb4-4e57-bce4-841ac7e24f23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.884725] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92be86d-be73-4052-9f19-6ba8d8a68c7b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.918840] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5c9669-d48b-4bad-a41f-eeb45347eaa1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.927723] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c1748e-7d53-49de-95bb-c63dd0714e96 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.943293] env[65788]: DEBUG nova.compute.provider_tree [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1278.103969] env[65788]: DEBUG oslo_concurrency.lockutils [None req-f1f6a469-7c5c-4017-bc4b-0a0bcff341d0 tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.056s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.222570] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663422, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.272503] env[65788]: DEBUG nova.compute.manager [req-e5072a0f-84ed-42a1-8ad7-82f2e5f98183 req-1d5430df-dcae-4144-942e-68d93fa95e8e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Received event network-vif-deleted-c6074f79-c6f2-4186-b2c3-809f045737ed {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1278.272709] env[65788]: INFO nova.compute.manager [req-e5072a0f-84ed-42a1-8ad7-82f2e5f98183 req-1d5430df-dcae-4144-942e-68d93fa95e8e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Neutron deleted interface c6074f79-c6f2-4186-b2c3-809f045737ed; detaching it from the instance and deleting it from the info cache [ 1278.272927] env[65788]: DEBUG nova.network.neutron [req-e5072a0f-84ed-42a1-8ad7-82f2e5f98183 req-1d5430df-dcae-4144-942e-68d93fa95e8e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1278.447263] env[65788]: DEBUG nova.scheduler.client.report [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1278.724107] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663422, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.746339] env[65788]: DEBUG nova.network.neutron [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1278.777055] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3529220f-480d-4c62-9993-90f91fc57b64 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.787318] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d1f0b8-4143-45d6-ad6a-dfaab402b78f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.822540] env[65788]: DEBUG nova.compute.manager [req-e5072a0f-84ed-42a1-8ad7-82f2e5f98183 req-1d5430df-dcae-4144-942e-68d93fa95e8e service nova] [instance: bdd09116-31de-491d-b129-c117f898881e] Detach interface failed, port_id=c6074f79-c6f2-4186-b2c3-809f045737ed, reason: Instance bdd09116-31de-491d-b129-c117f898881e could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1278.868184] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "16dfc544-45f3-486b-bc3d-86836382b07b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.868707] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.869266] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "16dfc544-45f3-486b-bc3d-86836382b07b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1278.869623] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1278.869872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.872845] env[65788]: INFO nova.compute.manager [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Terminating instance [ 1278.953072] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1278.983106] env[65788]: INFO nova.scheduler.client.report [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Deleted allocations for instance ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0 [ 1279.224925] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663422, 'name': ReconfigVM_Task, 'duration_secs': 1.784366} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.225332] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Reconfigured VM instance instance-0000007c to attach disk [datastore2] cc9be8c2-a6a8-4b33-9b15-e9ee58e43304/cc9be8c2-a6a8-4b33-9b15-e9ee58e43304.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1279.226034] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-315e6bbb-a6a5-4134-b0b0-406a6c868c1b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.234307] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1279.234307] env[65788]: value = "task-4663424" [ 1279.234307] env[65788]: _type = "Task" [ 1279.234307] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.244484] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663424, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.250089] env[65788]: INFO nova.compute.manager [-] [instance: bdd09116-31de-491d-b129-c117f898881e] Took 1.51 seconds to deallocate network for instance. [ 1279.378716] env[65788]: DEBUG nova.compute.manager [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1279.379058] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1279.380055] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cb9c6d-f0ab-4183-a9b5-856422020a83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.390095] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1279.390417] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c38f4c2c-5bf8-4b09-8543-d1b181766f1f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.463015] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1279.463295] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1279.463522] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore2] 16dfc544-45f3-486b-bc3d-86836382b07b {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1279.463785] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7490a046-e6f5-4550-a461-9b95dde23002 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.471579] env[65788]: DEBUG oslo_vmware.api [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1279.471579] env[65788]: value = "task-4663426" [ 1279.471579] env[65788]: _type = "Task" [ 1279.471579] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.480135] env[65788]: DEBUG oslo_vmware.api [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663426, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.493747] env[65788]: DEBUG oslo_concurrency.lockutils [None req-dd77db91-d9ae-4613-81f3-8a32c7eca660 tempest-AttachVolumeNegativeTest-1814624860 tempest-AttachVolumeNegativeTest-1814624860-project-member] Lock "ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.933s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1279.745315] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663424, 'name': Rename_Task, 'duration_secs': 0.218436} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.745594] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1279.745851] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26c63d06-8ba3-4ec4-a2aa-9317dfda59c8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.752385] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1279.752385] env[65788]: value = "task-4663427" [ 1279.752385] env[65788]: _type = "Task" [ 1279.752385] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1279.757351] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1279.757615] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1279.757821] env[65788]: DEBUG nova.objects.instance [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lazy-loading 'resources' on Instance uuid bdd09116-31de-491d-b129-c117f898881e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1279.764552] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663427, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.982514] env[65788]: DEBUG oslo_vmware.api [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663426, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137617} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.982779] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1279.982964] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1279.983168] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1279.983341] env[65788]: INFO nova.compute.manager [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1279.983594] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1279.983791] env[65788]: DEBUG nova.compute.manager [-] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1279.983903] env[65788]: DEBUG nova.network.neutron [-] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1279.984155] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1279.984675] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1279.984930] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1280.022668] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1280.265193] env[65788]: DEBUG oslo_vmware.api [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663427, 'name': PowerOnVM_Task, 'duration_secs': 0.454492} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.267701] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1280.267964] env[65788]: INFO nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Took 8.97 seconds to spawn the instance on the hypervisor. [ 1280.268191] env[65788]: DEBUG nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1280.269291] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94222f34-2543-4a0b-b83c-49a90f2c4185 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.296446] env[65788]: DEBUG nova.compute.manager [req-941169ad-32b0-4178-91db-e5c742536f08 req-6f2e8950-8aa3-4a2b-93f7-4bf534d5e809 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Received event network-vif-deleted-d7eca264-2c75-44a4-a968-213e84c89d5b {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1280.296647] env[65788]: INFO nova.compute.manager [req-941169ad-32b0-4178-91db-e5c742536f08 req-6f2e8950-8aa3-4a2b-93f7-4bf534d5e809 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Neutron deleted interface d7eca264-2c75-44a4-a968-213e84c89d5b; detaching it from the instance and deleting it from the info cache [ 1280.296822] env[65788]: DEBUG nova.network.neutron [req-941169ad-32b0-4178-91db-e5c742536f08 req-6f2e8950-8aa3-4a2b-93f7-4bf534d5e809 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1280.381042] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767b6478-71e2-4d3d-919d-5c0099b5b247 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.389345] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31dae310-8dee-43b4-843b-8f8bcc41318a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.419850] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4d09e9-121c-4bfe-8f03-af5570e0e19e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.428050] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cd145e-287c-4c34-8462-8e4b8d06fe4a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.441929] env[65788]: DEBUG nova.compute.provider_tree [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.769509] env[65788]: DEBUG nova.network.neutron [-] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1280.787014] env[65788]: INFO nova.compute.manager [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Took 13.75 seconds to build instance. [ 1280.799216] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5e0e1da-db93-44d2-9bb8-8a8e98018866 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.809630] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d917876e-acb6-4388-aa06-8ff2d03e59a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.841095] env[65788]: DEBUG nova.compute.manager [req-941169ad-32b0-4178-91db-e5c742536f08 req-6f2e8950-8aa3-4a2b-93f7-4bf534d5e809 service nova] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Detach interface failed, port_id=d7eca264-2c75-44a4-a968-213e84c89d5b, reason: Instance 16dfc544-45f3-486b-bc3d-86836382b07b could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1280.944887] env[65788]: DEBUG nova.scheduler.client.report [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1281.272878] env[65788]: INFO nova.compute.manager [-] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Took 1.29 seconds to deallocate network for instance. [ 1281.289397] env[65788]: DEBUG oslo_concurrency.lockutils [None req-7358ed4e-3322-4de4-b437-848cb7b163cd tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.257s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1281.449828] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1281.478318] env[65788]: INFO nova.scheduler.client.report [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Deleted allocations for instance bdd09116-31de-491d-b129-c117f898881e [ 1281.744118] env[65788]: DEBUG nova.compute.manager [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1281.744118] env[65788]: DEBUG nova.compute.manager [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing instance network info cache due to event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1281.744118] env[65788]: DEBUG oslo_concurrency.lockutils [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.744118] env[65788]: DEBUG oslo_concurrency.lockutils [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1281.744118] env[65788]: DEBUG nova.network.neutron [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1281.779719] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1281.780031] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1281.780272] env[65788]: DEBUG nova.objects.instance [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid 16dfc544-45f3-486b-bc3d-86836382b07b {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1281.984879] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2bb91d66-1284-48a3-8fb3-080a49bd4888 tempest-AttachVolumeShelveTestJSON-903356803 tempest-AttachVolumeShelveTestJSON-903356803-project-member] Lock "bdd09116-31de-491d-b129-c117f898881e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.889s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1282.246017] env[65788]: WARNING neutronclient.v2_0.client [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1282.246818] env[65788]: WARNING openstack [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1282.247225] env[65788]: WARNING openstack [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.434227] env[65788]: WARNING openstack [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1282.434640] env[65788]: WARNING openstack [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.479383] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101f866e-d00d-4b3a-8307-78b5563b308a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.488751] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2fbe22-bd21-4f96-89ac-9ce714efd453 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.524702] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43bf89b6-2206-4061-a30f-2e874c7eec4d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.534154] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aad0a5f-c162-44f3-a7d0-37145a2fbac9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.549772] env[65788]: DEBUG nova.compute.provider_tree [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1282.707995] env[65788]: WARNING neutronclient.v2_0.client [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1282.708686] env[65788]: WARNING openstack [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1282.709035] env[65788]: WARNING openstack [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1282.796617] env[65788]: DEBUG nova.network.neutron [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updated VIF entry in instance network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1282.797192] env[65788]: DEBUG nova.network.neutron [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1282.883423] env[65788]: DEBUG nova.compute.manager [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1282.883620] env[65788]: DEBUG nova.compute.manager [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing instance network info cache due to event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1282.883821] env[65788]: DEBUG oslo_concurrency.lockutils [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.883954] env[65788]: DEBUG oslo_concurrency.lockutils [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1282.884436] env[65788]: DEBUG nova.network.neutron [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1283.054096] env[65788]: DEBUG nova.scheduler.client.report [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1283.300190] env[65788]: DEBUG oslo_concurrency.lockutils [req-45f5fe95-f263-4aa3-972c-be723cffa2ad req-8fcca043-f557-4126-a061-73c050caf597 service nova] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1283.387826] env[65788]: WARNING neutronclient.v2_0.client [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.388673] env[65788]: WARNING openstack [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.389086] env[65788]: WARNING openstack [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.492855] env[65788]: WARNING openstack [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.493643] env[65788]: WARNING openstack [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.553747] env[65788]: WARNING neutronclient.v2_0.client [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1283.554454] env[65788]: WARNING openstack [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1283.554838] env[65788]: WARNING openstack [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1283.562574] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.783s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1283.591801] env[65788]: INFO nova.scheduler.client.report [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance 16dfc544-45f3-486b-bc3d-86836382b07b [ 1283.642284] env[65788]: DEBUG nova.network.neutron [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updated VIF entry in instance network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1283.642650] env[65788]: DEBUG nova.network.neutron [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1283.778578] env[65788]: DEBUG nova.compute.manager [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1283.778578] env[65788]: DEBUG nova.compute.manager [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing instance network info cache due to event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1283.778578] env[65788]: DEBUG oslo_concurrency.lockutils [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.100888] env[65788]: DEBUG oslo_concurrency.lockutils [None req-ad33ab64-076e-47db-9270-ab4d3302128a tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "16dfc544-45f3-486b-bc3d-86836382b07b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.232s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1284.145356] env[65788]: DEBUG oslo_concurrency.lockutils [req-ae2e62ce-bbe9-4d18-86af-8d81082322f6 req-7a818be8-4352-4bef-9780-4e4055393f19 service nova] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.146254] env[65788]: DEBUG oslo_concurrency.lockutils [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1284.146533] env[65788]: DEBUG nova.network.neutron [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1284.649263] env[65788]: WARNING neutronclient.v2_0.client [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1284.649978] env[65788]: WARNING openstack [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1284.650363] env[65788]: WARNING openstack [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.745165] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1284.745440] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1284.745645] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "27ed2344-c42b-46bc-b51a-20821f67cbf0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1284.745828] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1284.745991] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1284.748701] env[65788]: WARNING openstack [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1284.749160] env[65788]: WARNING openstack [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.755848] env[65788]: INFO nova.compute.manager [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Terminating instance [ 1284.809088] env[65788]: WARNING neutronclient.v2_0.client [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1284.809754] env[65788]: WARNING openstack [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1284.810116] env[65788]: WARNING openstack [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1284.889426] env[65788]: DEBUG nova.network.neutron [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updated VIF entry in instance network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1284.889788] env[65788]: DEBUG nova.network.neutron [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1284.911146] env[65788]: DEBUG nova.compute.manager [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1284.911146] env[65788]: DEBUG nova.compute.manager [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing instance network info cache due to event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1284.911249] env[65788]: DEBUG oslo_concurrency.lockutils [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.911315] env[65788]: DEBUG oslo_concurrency.lockutils [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1284.911476] env[65788]: DEBUG nova.network.neutron [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1285.260209] env[65788]: DEBUG nova.compute.manager [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1285.260448] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1285.261374] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8ee2f1-4e9e-4c7e-83c0-bf70e1f809a8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.269911] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1285.270170] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-997fad54-c4e4-4be6-a2e1-a93ebe3467a9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.276981] env[65788]: DEBUG oslo_vmware.api [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1285.276981] env[65788]: value = "task-4663431" [ 1285.276981] env[65788]: _type = "Task" [ 1285.276981] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.284909] env[65788]: DEBUG oslo_vmware.api [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663431, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.394007] env[65788]: DEBUG oslo_concurrency.lockutils [req-d8aa4940-4bff-4f71-b3f3-b6b013f7c54b req-b2f61633-9054-4580-871b-bec3bf2e9e17 service nova] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1285.414893] env[65788]: WARNING neutronclient.v2_0.client [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1285.415810] env[65788]: WARNING openstack [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.416208] env[65788]: WARNING openstack [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1285.549453] env[65788]: WARNING openstack [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.549903] env[65788]: WARNING openstack [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1285.608397] env[65788]: WARNING neutronclient.v2_0.client [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1285.609130] env[65788]: WARNING openstack [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1285.609487] env[65788]: WARNING openstack [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1285.688998] env[65788]: DEBUG nova.network.neutron [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updated VIF entry in instance network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1285.689396] env[65788]: DEBUG nova.network.neutron [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1285.787717] env[65788]: DEBUG oslo_vmware.api [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663431, 'name': PowerOffVM_Task, 'duration_secs': 0.19049} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.787976] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1285.788128] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1285.788399] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7c9ebed-e348-47e5-994d-90f8617badec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.859946] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1285.860201] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Deleting contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1285.860400] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleting the datastore file [datastore1] 27ed2344-c42b-46bc-b51a-20821f67cbf0 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1285.860690] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93a9500f-b64c-44f5-8981-16af1312c571 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.868998] env[65788]: DEBUG oslo_vmware.api [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for the task: (returnval){ [ 1285.868998] env[65788]: value = "task-4663433" [ 1285.868998] env[65788]: _type = "Task" [ 1285.868998] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.877309] env[65788]: DEBUG oslo_vmware.api [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663433, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.192742] env[65788]: DEBUG oslo_concurrency.lockutils [req-6e52b103-3692-4840-95dc-9f60479276c7 req-4a28afe6-6e77-43d0-9719-6390428dfb50 service nova] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1286.379828] env[65788]: DEBUG oslo_vmware.api [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Task: {'id': task-4663433, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134302} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.380102] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1286.380290] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Deleted contents of the VM from datastore datastore1 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1286.380464] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1286.380632] env[65788]: INFO nova.compute.manager [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1286.380899] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1286.381117] env[65788]: DEBUG nova.compute.manager [-] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1286.381306] env[65788]: DEBUG nova.network.neutron [-] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1286.381555] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.382082] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1286.382345] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1286.445134] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1286.676976] env[65788]: DEBUG nova.compute.manager [req-f76b1bdb-f414-4893-b611-edfc0d7b0e4e req-eb041300-b2b0-4b7a-b344-4ee5e5cd66bf service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Received event network-vif-deleted-c2856b60-c6f3-40b5-a301-a764e0b447eb {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1286.677210] env[65788]: INFO nova.compute.manager [req-f76b1bdb-f414-4893-b611-edfc0d7b0e4e req-eb041300-b2b0-4b7a-b344-4ee5e5cd66bf service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Neutron deleted interface c2856b60-c6f3-40b5-a301-a764e0b447eb; detaching it from the instance and deleting it from the info cache [ 1286.677413] env[65788]: DEBUG nova.network.neutron [req-f76b1bdb-f414-4893-b611-edfc0d7b0e4e req-eb041300-b2b0-4b7a-b344-4ee5e5cd66bf service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1287.147144] env[65788]: DEBUG nova.network.neutron [-] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1287.181294] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd24cedb-7841-45dc-959d-3bdc2fb8b020 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.193101] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0663fac8-7b04-4f04-969b-b6ac64fc3314 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.224650] env[65788]: DEBUG nova.compute.manager [req-f76b1bdb-f414-4893-b611-edfc0d7b0e4e req-eb041300-b2b0-4b7a-b344-4ee5e5cd66bf service nova] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Detach interface failed, port_id=c2856b60-c6f3-40b5-a301-a764e0b447eb, reason: Instance 27ed2344-c42b-46bc-b51a-20821f67cbf0 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1287.649447] env[65788]: INFO nova.compute.manager [-] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Took 1.27 seconds to deallocate network for instance. [ 1288.156954] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1288.157400] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1288.157400] env[65788]: DEBUG nova.objects.instance [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lazy-loading 'resources' on Instance uuid 27ed2344-c42b-46bc-b51a-20821f67cbf0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1288.735522] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd267712-2ab6-4786-9b76-b580d4d031b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.744415] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6317b5a9-5f8d-4968-957c-76cdab554350 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.775382] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1a999c-c454-483f-8f01-09ee45bf2d71 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.784235] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808d1004-bc2f-4a9d-abbb-30a9d3ebbd43 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.799482] env[65788]: DEBUG nova.compute.provider_tree [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1289.303444] env[65788]: DEBUG nova.scheduler.client.report [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1289.809444] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.652s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.834118] env[65788]: INFO nova.scheduler.client.report [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Deleted allocations for instance 27ed2344-c42b-46bc-b51a-20821f67cbf0 [ 1290.344213] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d5e226d1-1ca1-4289-8b33-9bca613d3c2f tempest-ServersTestJSON-581224055 tempest-ServersTestJSON-581224055-project-member] Lock "27ed2344-c42b-46bc-b51a-20821f67cbf0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.599s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1297.154116] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-4abbb248-12f6-4612-b6ea-01d1ef2104d7-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1297.154461] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-4abbb248-12f6-4612-b6ea-01d1ef2104d7-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1297.154830] env[65788]: DEBUG nova.objects.instance [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'flavor' on Instance uuid 4abbb248-12f6-4612-b6ea-01d1ef2104d7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.658969] env[65788]: WARNING neutronclient.v2_0.client [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1297.659716] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1297.660121] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1297.765761] env[65788]: DEBUG nova.objects.instance [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'pci_requests' on Instance uuid 4abbb248-12f6-4612-b6ea-01d1ef2104d7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1298.268924] env[65788]: DEBUG nova.objects.base [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Object Instance<4abbb248-12f6-4612-b6ea-01d1ef2104d7> lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1298.269323] env[65788]: DEBUG nova.network.neutron [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1298.269504] env[65788]: WARNING neutronclient.v2_0.client [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1298.269800] env[65788]: WARNING neutronclient.v2_0.client [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1298.270377] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1298.270715] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1298.340046] env[65788]: DEBUG nova.policy [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1299.719929] env[65788]: DEBUG nova.compute.manager [req-fa729701-cd28-46d9-b554-54f646e0b82e req-a7be69ac-3c56-47b2-9134-b8e538eb9831 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-vif-plugged-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1299.720242] env[65788]: DEBUG oslo_concurrency.lockutils [req-fa729701-cd28-46d9-b554-54f646e0b82e req-a7be69ac-3c56-47b2-9134-b8e538eb9831 service nova] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1299.720363] env[65788]: DEBUG oslo_concurrency.lockutils [req-fa729701-cd28-46d9-b554-54f646e0b82e req-a7be69ac-3c56-47b2-9134-b8e538eb9831 service nova] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1299.720628] env[65788]: DEBUG oslo_concurrency.lockutils [req-fa729701-cd28-46d9-b554-54f646e0b82e req-a7be69ac-3c56-47b2-9134-b8e538eb9831 service nova] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1299.720697] env[65788]: DEBUG nova.compute.manager [req-fa729701-cd28-46d9-b554-54f646e0b82e req-a7be69ac-3c56-47b2-9134-b8e538eb9831 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] No waiting events found dispatching network-vif-plugged-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1299.720853] env[65788]: WARNING nova.compute.manager [req-fa729701-cd28-46d9-b554-54f646e0b82e req-a7be69ac-3c56-47b2-9134-b8e538eb9831 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received unexpected event network-vif-plugged-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 for instance with vm_state active and task_state None. [ 1299.794492] env[65788]: DEBUG nova.network.neutron [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Successfully updated port: e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1300.297657] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1300.297657] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1300.297892] env[65788]: DEBUG nova.network.neutron [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1300.800871] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1300.801306] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1300.839915] env[65788]: WARNING nova.network.neutron [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1300.877018] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1300.877425] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1300.936269] env[65788]: WARNING neutronclient.v2_0.client [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1300.936941] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1300.937297] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1301.035524] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1301.035932] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1301.093887] env[65788]: WARNING neutronclient.v2_0.client [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1301.094574] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1301.094910] env[65788]: WARNING openstack [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1301.174279] env[65788]: DEBUG nova.network.neutron [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "address": "fa:16:3e:44:58:73", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0ea25a5-c2", "ovs_interfaceid": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1301.677416] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1301.678110] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.678276] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1301.679154] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8421ede0-487a-40c7-85bd-dd8c333a30dc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.697842] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1301.698220] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1301.698487] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1301.698797] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1301.699067] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1301.699346] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1301.699693] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1301.699958] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1301.700256] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1301.700550] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1301.700837] env[65788]: DEBUG nova.virt.hardware [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1301.707723] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Reconfiguring VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1301.708076] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c37841c3-3b25-4961-95a0-2afa0b675ecd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.726096] env[65788]: DEBUG oslo_vmware.api [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1301.726096] env[65788]: value = "task-4663435" [ 1301.726096] env[65788]: _type = "Task" [ 1301.726096] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.734515] env[65788]: DEBUG oslo_vmware.api [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663435, 'name': ReconfigVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.787155] env[65788]: DEBUG nova.compute.manager [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-changed-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1301.787372] env[65788]: DEBUG nova.compute.manager [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing instance network info cache due to event network-changed-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1301.787710] env[65788]: DEBUG oslo_concurrency.lockutils [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.787802] env[65788]: DEBUG oslo_concurrency.lockutils [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1301.787908] env[65788]: DEBUG nova.network.neutron [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing network info cache for port e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1302.236721] env[65788]: DEBUG oslo_vmware.api [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663435, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.290641] env[65788]: WARNING neutronclient.v2_0.client [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1302.291408] env[65788]: WARNING openstack [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1302.291768] env[65788]: WARNING openstack [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1302.406065] env[65788]: WARNING openstack [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1302.406546] env[65788]: WARNING openstack [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1302.465212] env[65788]: WARNING neutronclient.v2_0.client [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1302.465918] env[65788]: WARNING openstack [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1302.466312] env[65788]: WARNING openstack [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1302.548082] env[65788]: DEBUG nova.network.neutron [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updated VIF entry in instance network info cache for port e0ea25a5-c2ed-405d-b6bf-bc16e70fd558. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1302.548527] env[65788]: DEBUG nova.network.neutron [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "address": "fa:16:3e:44:58:73", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0ea25a5-c2", "ovs_interfaceid": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1302.737041] env[65788]: DEBUG oslo_vmware.api [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663435, 'name': ReconfigVM_Task, 'duration_secs': 0.587162} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.737421] env[65788]: WARNING neutronclient.v2_0.client [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1302.737662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1302.737924] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Reconfigured VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1303.051337] env[65788]: DEBUG oslo_concurrency.lockutils [req-0ce8712d-66b6-4275-be70-9e917d4c66cd req-0256240d-a3b2-45ef-a6ec-203c567c914d service nova] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1303.242204] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c358ad06-b378-46a7-84ec-cd3aa08bb743 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-4abbb248-12f6-4612-b6ea-01d1ef2104d7-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.088s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1304.873064] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-4abbb248-12f6-4612-b6ea-01d1ef2104d7-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1304.873495] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-4abbb248-12f6-4612-b6ea-01d1ef2104d7-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1305.375859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1305.376077] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1305.377015] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d0f517-9d1f-45a2-b1d0-3f63267b0703 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.395940] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc0e197-c36d-454a-a60e-3c4ca10369f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.417290] env[65788]: WARNING neutronclient.v2_0.client [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1305.422666] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Reconfiguring VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1305.422959] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8fe74e5-f11d-416f-8a9c-74f278c34601 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.441888] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1305.441888] env[65788]: value = "task-4663436" [ 1305.441888] env[65788]: _type = "Task" [ 1305.441888] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.450243] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.952195] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.452693] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.794727] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "222d1456-baad-42ee-bdef-020fb66ebe41" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.795139] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "222d1456-baad-42ee-bdef-020fb66ebe41" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.795439] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "222d1456-baad-42ee-bdef-020fb66ebe41-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.795641] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "222d1456-baad-42ee-bdef-020fb66ebe41-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.795812] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "222d1456-baad-42ee-bdef-020fb66ebe41-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.799523] env[65788]: INFO nova.compute.manager [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Terminating instance [ 1306.953225] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.304096] env[65788]: DEBUG nova.compute.manager [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1307.304316] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1307.305286] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baead8eb-49ef-4a1e-aa97-a31895f052f1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.313086] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1307.313345] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c256e33a-d7e5-4321-a7d4-74d894a4d5d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.319931] env[65788]: DEBUG oslo_vmware.api [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1307.319931] env[65788]: value = "task-4663437" [ 1307.319931] env[65788]: _type = "Task" [ 1307.319931] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.328961] env[65788]: DEBUG oslo_vmware.api [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.456629] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.830163] env[65788]: DEBUG oslo_vmware.api [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663437, 'name': PowerOffVM_Task, 'duration_secs': 0.234912} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.830464] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1307.830580] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1307.830847] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-024c7a59-ebfd-44e8-955a-8f4a7600d6f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.898628] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1307.898914] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1307.899155] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleting the datastore file [datastore2] 222d1456-baad-42ee-bdef-020fb66ebe41 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1307.899489] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10e141c2-6135-4bb1-a596-303822510c92 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.906893] env[65788]: DEBUG oslo_vmware.api [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1307.906893] env[65788]: value = "task-4663439" [ 1307.906893] env[65788]: _type = "Task" [ 1307.906893] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.914828] env[65788]: DEBUG oslo_vmware.api [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663439, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.954533] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.417405] env[65788]: DEBUG oslo_vmware.api [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663439, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147357} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.417672] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1308.417885] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1308.418095] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1308.418275] env[65788]: INFO nova.compute.manager [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1308.418541] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1308.418733] env[65788]: DEBUG nova.compute.manager [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1308.418832] env[65788]: DEBUG nova.network.neutron [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1308.419084] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.419614] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1308.419903] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1308.455472] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.457401] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1308.929035] env[65788]: DEBUG nova.compute.manager [req-5f52e812-f3cf-4ea5-be83-a9e4f41b445c req-8a5d8a2d-c8fe-4c4f-bb5d-b980a800496a service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Received event network-vif-deleted-fff8c844-a1fe-4657-9ab0-99fa3a35d346 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1308.929230] env[65788]: INFO nova.compute.manager [req-5f52e812-f3cf-4ea5-be83-a9e4f41b445c req-8a5d8a2d-c8fe-4c4f-bb5d-b980a800496a service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Neutron deleted interface fff8c844-a1fe-4657-9ab0-99fa3a35d346; detaching it from the instance and deleting it from the info cache [ 1308.929349] env[65788]: DEBUG nova.network.neutron [req-5f52e812-f3cf-4ea5-be83-a9e4f41b445c req-8a5d8a2d-c8fe-4c4f-bb5d-b980a800496a service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1308.955619] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.401547] env[65788]: DEBUG nova.network.neutron [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1309.432410] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b659814-a104-43c5-8326-81044926b971 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.443189] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e61a90-b49b-4daf-9008-c15359321b93 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.463409] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.473825] env[65788]: DEBUG nova.compute.manager [req-5f52e812-f3cf-4ea5-be83-a9e4f41b445c req-8a5d8a2d-c8fe-4c4f-bb5d-b980a800496a service nova] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Detach interface failed, port_id=fff8c844-a1fe-4657-9ab0-99fa3a35d346, reason: Instance 222d1456-baad-42ee-bdef-020fb66ebe41 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1309.904855] env[65788]: INFO nova.compute.manager [-] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Took 1.49 seconds to deallocate network for instance. [ 1309.964625] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.411456] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1310.411704] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1310.411925] env[65788]: DEBUG nova.objects.instance [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'resources' on Instance uuid 222d1456-baad-42ee-bdef-020fb66ebe41 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1310.465157] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.967913] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.982896] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3c836a-06b1-4705-9226-daa10ade37b3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.991016] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99fdbf4-ba34-4ece-b186-87beb9dfd274 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.022873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f59506-29e3-4427-a45d-4d8b1e6047cc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.030835] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dcc809c-1f87-42a3-af8e-c75652c17277 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.045866] env[65788]: DEBUG nova.compute.provider_tree [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1311.466211] env[65788]: DEBUG oslo_vmware.api [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663436, 'name': ReconfigVM_Task, 'duration_secs': 5.766798} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.466503] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1311.466707] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Reconfigured VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1311.467140] env[65788]: WARNING neutronclient.v2_0.client [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.467457] env[65788]: WARNING neutronclient.v2_0.client [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.468019] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1311.468357] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1311.501845] env[65788]: WARNING neutronclient.v2_0.client [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1311.548823] env[65788]: DEBUG nova.scheduler.client.report [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1312.053679] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.642s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.074296] env[65788]: INFO nova.scheduler.client.report [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted allocations for instance 222d1456-baad-42ee-bdef-020fb66ebe41 [ 1312.582915] env[65788]: DEBUG oslo_concurrency.lockutils [None req-6756794c-a844-4cc6-a400-152ba6e96d61 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "222d1456-baad-42ee-bdef-020fb66ebe41" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.788s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.797819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1312.798027] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1312.798211] env[65788]: DEBUG nova.network.neutron [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1313.301241] env[65788]: WARNING neutronclient.v2_0.client [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1313.301961] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1313.302609] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1313.410736] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1313.411188] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1313.479025] env[65788]: WARNING neutronclient.v2_0.client [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1313.479714] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1313.480078] env[65788]: WARNING openstack [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1313.505192] env[65788]: DEBUG nova.compute.manager [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1313.505313] env[65788]: DEBUG nova.compute.manager [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing instance network info cache due to event network-changed-0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1313.505460] env[65788]: DEBUG oslo_concurrency.lockutils [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Acquiring lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1313.579130] env[65788]: INFO nova.network.neutron [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Port e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1313.579130] env[65788]: DEBUG nova.network.neutron [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1314.082329] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1314.084571] env[65788]: DEBUG oslo_concurrency.lockutils [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Acquired lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1314.084717] env[65788]: DEBUG nova.network.neutron [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Refreshing network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1314.183997] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1314.184312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1314.185057] env[65788]: DEBUG nova.objects.instance [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'flavor' on Instance uuid cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1314.588457] env[65788]: WARNING neutronclient.v2_0.client [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1314.589135] env[65788]: WARNING openstack [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.589471] env[65788]: WARNING openstack [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.596837] env[65788]: DEBUG oslo_concurrency.lockutils [None req-2d2356d4-f9b4-4d7e-8de0-231282381a8b tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-4abbb248-12f6-4612-b6ea-01d1ef2104d7-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.723s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1314.688563] env[65788]: WARNING neutronclient.v2_0.client [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1314.689246] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.689583] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.699138] env[65788]: WARNING openstack [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.699486] env[65788]: WARNING openstack [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.768340] env[65788]: WARNING neutronclient.v2_0.client [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1314.769185] env[65788]: WARNING openstack [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1314.769548] env[65788]: WARNING openstack [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1314.801167] env[65788]: DEBUG nova.objects.instance [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'pci_requests' on Instance uuid cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1314.808044] env[65788]: DEBUG nova.compute.manager [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1314.852288] env[65788]: DEBUG nova.network.neutron [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updated VIF entry in instance network info cache for port 0c1d654b-57fe-46a4-a1f9-63029c094c97. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1314.852671] env[65788]: DEBUG nova.network.neutron [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [{"id": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "address": "fa:16:3e:02:4d:d5", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c1d654b-57", "ovs_interfaceid": "0c1d654b-57fe-46a4-a1f9-63029c094c97", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1314.905943] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1314.906191] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1314.906384] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1314.906546] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.304148] env[65788]: DEBUG nova.objects.base [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1315.304375] env[65788]: DEBUG nova.network.neutron [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1315.304731] env[65788]: WARNING neutronclient.v2_0.client [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1315.305050] env[65788]: WARNING neutronclient.v2_0.client [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1315.305699] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1315.306062] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1315.333009] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1315.333351] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1315.355123] env[65788]: DEBUG oslo_concurrency.lockutils [req-da996a39-9c98-4b9a-910c-9352383848dd req-04784772-2580-4976-9d0e-adba9c11e3a9 service nova] Releasing lock "refresh_cache-4abbb248-12f6-4612-b6ea-01d1ef2104d7" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1315.397862] env[65788]: DEBUG nova.policy [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6beba3c760f543a9b577672ffc714ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676963c25a3741998b10790689242dbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1315.536693] env[65788]: DEBUG nova.compute.manager [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1315.536906] env[65788]: DEBUG nova.compute.manager [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing instance network info cache due to event network-changed-b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1315.537145] env[65788]: DEBUG oslo_concurrency.lockutils [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1315.537272] env[65788]: DEBUG oslo_concurrency.lockutils [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1315.537431] env[65788]: DEBUG nova.network.neutron [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1315.638499] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1315.838920] env[65788]: INFO nova.compute.claims [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1316.040147] env[65788]: WARNING neutronclient.v2_0.client [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1316.040897] env[65788]: WARNING openstack [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.041264] env[65788]: WARNING openstack [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.170610] env[65788]: WARNING openstack [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.171048] env[65788]: WARNING openstack [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.234874] env[65788]: WARNING neutronclient.v2_0.client [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1316.235614] env[65788]: WARNING openstack [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1316.235964] env[65788]: WARNING openstack [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1316.318202] env[65788]: DEBUG nova.network.neutron [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updated VIF entry in instance network info cache for port b6e00388-271a-41b9-bb29-4463d4a692ea. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1316.318603] env[65788]: DEBUG nova.network.neutron [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1316.345326] env[65788]: INFO nova.compute.resource_tracker [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating resource usage from migration ba8aa17b-96d2-40b6-8a39-4e67c1e48856 [ 1316.412070] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b017e874-846d-4b93-bf57-46fa8bc53cb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.420281] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce8a4cb-8a19-4437-8f04-402466a4872e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.451852] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15152f4-9409-426c-b259-fbfb3e5f4b0e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.460120] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc52e7f-48a2-4782-91d2-3985b3ca7bed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.474028] env[65788]: DEBUG nova.compute.provider_tree [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1316.637617] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1316.637824] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1316.775728] env[65788]: DEBUG nova.compute.manager [req-6c734ae6-87fd-42fc-9cd3-a2a994deb5da req-03069de2-eb3f-4774-96cd-e70a035f45bd service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-vif-plugged-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1316.776047] env[65788]: DEBUG oslo_concurrency.lockutils [req-6c734ae6-87fd-42fc-9cd3-a2a994deb5da req-03069de2-eb3f-4774-96cd-e70a035f45bd service nova] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1316.776351] env[65788]: DEBUG oslo_concurrency.lockutils [req-6c734ae6-87fd-42fc-9cd3-a2a994deb5da req-03069de2-eb3f-4774-96cd-e70a035f45bd service nova] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1316.776593] env[65788]: DEBUG oslo_concurrency.lockutils [req-6c734ae6-87fd-42fc-9cd3-a2a994deb5da req-03069de2-eb3f-4774-96cd-e70a035f45bd service nova] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1316.776777] env[65788]: DEBUG nova.compute.manager [req-6c734ae6-87fd-42fc-9cd3-a2a994deb5da req-03069de2-eb3f-4774-96cd-e70a035f45bd service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] No waiting events found dispatching network-vif-plugged-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1316.776938] env[65788]: WARNING nova.compute.manager [req-6c734ae6-87fd-42fc-9cd3-a2a994deb5da req-03069de2-eb3f-4774-96cd-e70a035f45bd service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received unexpected event network-vif-plugged-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 for instance with vm_state active and task_state None. [ 1316.821699] env[65788]: DEBUG oslo_concurrency.lockutils [req-298245a7-3ee2-4519-9d5e-12ed31d6ce68 req-9de421d7-1e97-4a26-b3ef-a430c55ce0c3 service nova] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1316.855415] env[65788]: DEBUG nova.network.neutron [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Successfully updated port: e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1316.977583] env[65788]: DEBUG nova.scheduler.client.report [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1317.358131] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1317.358342] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1317.358512] env[65788]: DEBUG nova.network.neutron [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1317.483334] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.150s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1317.483575] env[65788]: INFO nova.compute.manager [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Migrating [ 1317.632391] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1317.861590] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1317.862029] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1317.900734] env[65788]: WARNING nova.network.neutron [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] b03fb885-c6ac-4708-9401-d5b501cae296 already exists in list: networks containing: ['b03fb885-c6ac-4708-9401-d5b501cae296']. ignoring it [ 1317.937497] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1317.937884] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.000526] env[65788]: WARNING neutronclient.v2_0.client [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1318.001216] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1318.001608] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.009346] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.009511] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1318.009686] env[65788]: DEBUG nova.network.neutron [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1318.113627] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1318.114031] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.176843] env[65788]: WARNING neutronclient.v2_0.client [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1318.177731] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1318.178138] env[65788]: WARNING openstack [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.257175] env[65788]: DEBUG nova.network.neutron [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "address": "fa:16:3e:44:58:73", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0ea25a5-c2", "ovs_interfaceid": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1318.512981] env[65788]: WARNING neutronclient.v2_0.client [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1318.513689] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1318.514050] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.626639] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1318.627042] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.637700] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.686763] env[65788]: WARNING neutronclient.v2_0.client [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1318.687450] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1318.687847] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1318.759677] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1318.760357] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.760533] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1318.761395] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9b7ad0-2b7e-4c25-b14c-f9dd45b6d477 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.767497] env[65788]: DEBUG nova.network.neutron [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1318.782240] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1318.782447] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1318.782621] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1318.782764] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1318.782904] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1318.783064] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1318.783285] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1318.783442] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1318.783674] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1318.783942] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1318.784068] env[65788]: DEBUG nova.virt.hardware [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1318.790260] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Reconfiguring VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1318.790957] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1318.792393] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-021d5e4a-259a-4abc-8bca-0322a30c7dba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.811530] env[65788]: DEBUG nova.compute.manager [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-changed-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1318.811757] env[65788]: DEBUG nova.compute.manager [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing instance network info cache due to event network-changed-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1318.811918] env[65788]: DEBUG oslo_concurrency.lockutils [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1318.812068] env[65788]: DEBUG oslo_concurrency.lockutils [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1318.812224] env[65788]: DEBUG nova.network.neutron [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Refreshing network info cache for port e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1318.816446] env[65788]: DEBUG oslo_vmware.api [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1318.816446] env[65788]: value = "task-4663440" [ 1318.816446] env[65788]: _type = "Task" [ 1318.816446] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1318.825973] env[65788]: DEBUG oslo_vmware.api [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663440, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.140820] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.141109] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.141264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1319.141420] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1319.142721] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef571f2b-4c69-43c0-986a-19669b191bb2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.151372] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5b169e-dfbb-4259-8025-e92c79628d7c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.165831] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a544057-cc9d-40f2-a11f-85e3bad9875f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.174077] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6ecace-1ede-4560-877b-e24312a8aecf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.205810] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179829MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1319.206079] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1319.206232] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1319.314384] env[65788]: WARNING neutronclient.v2_0.client [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1319.315219] env[65788]: WARNING openstack [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1319.315669] env[65788]: WARNING openstack [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1319.334219] env[65788]: DEBUG oslo_vmware.api [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663440, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1319.440051] env[65788]: WARNING openstack [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1319.440536] env[65788]: WARNING openstack [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1319.502858] env[65788]: WARNING neutronclient.v2_0.client [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1319.503546] env[65788]: WARNING openstack [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1319.503940] env[65788]: WARNING openstack [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1319.582363] env[65788]: DEBUG nova.network.neutron [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updated VIF entry in instance network info cache for port e0ea25a5-c2ed-405d-b6bf-bc16e70fd558. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1319.582760] env[65788]: DEBUG nova.network.neutron [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "address": "fa:16:3e:44:58:73", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0ea25a5-c2", "ovs_interfaceid": "e0ea25a5-c2ed-405d-b6bf-bc16e70fd558", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1319.817328] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d58ad28-e726-4cc0-b54b-4b945be2ac83 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1319.835975] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1319.847931] env[65788]: DEBUG oslo_vmware.api [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663440, 'name': ReconfigVM_Task, 'duration_secs': 0.569226} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1319.848355] env[65788]: WARNING neutronclient.v2_0.client [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1319.848589] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1319.848794] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Reconfigured VM to attach interface {{(pid=65788) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1320.086224] env[65788]: DEBUG oslo_concurrency.lockutils [req-641981ba-5d5e-4c25-a712-0d430dc37368 req-6eeb52e0-b870-45b2-8113-93319aafba4e service nova] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1320.213810] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Applying migration context for instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e as it has an incoming, in-progress migration ba8aa17b-96d2-40b6-8a39-4e67c1e48856. Migration status is migrating {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 1320.214599] env[65788]: INFO nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating resource usage from migration ba8aa17b-96d2-40b6-8a39-4e67c1e48856 [ 1320.234204] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 4abbb248-12f6-4612-b6ea-01d1ef2104d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1320.234361] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1320.234484] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Migration ba8aa17b-96d2-40b6-8a39-4e67c1e48856 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 1320.234604] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1320.234783] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1320.234929] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '3', 'num_vm_active': '3', 'num_task_resize_migrating': '1', 'num_os_type_None': '3', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '1', 'io_workload': '1', 'num_task_None': '2', 'num_proj_676963c25a3741998b10790689242dbb': '2'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1320.301527] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c66c6c-e248-4bff-be5c-a687addb02ea {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.310660] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127f8152-c1b4-4591-ba6f-dc4a4d457a2c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.342214] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fab19e-dc0a-4268-9771-a45e98258ee9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.346150] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1320.346478] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca9fe4d0-338d-4562-b5b0-4a7385b1730b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.353787] env[65788]: DEBUG oslo_concurrency.lockutils [None req-25875925-d2b7-49e0-a163-70f72fb0c304 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.169s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1320.356886] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc2e0c7-e642-41cb-9927-aa9c4dad1e80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.361216] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1320.361216] env[65788]: value = "task-4663441" [ 1320.361216] env[65788]: _type = "Task" [ 1320.361216] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1320.374012] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1320.382111] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663441, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1320.876449] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663441, 'name': PowerOffVM_Task, 'duration_secs': 0.23921} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1320.877365] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1320.881439] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1320.881839] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1321.383217] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1321.383497] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.177s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1321.389027] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1321.389306] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1321.389603] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1321.389768] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1321.389921] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1321.390084] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1321.390298] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1321.390488] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1321.390660] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1321.390823] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1321.390994] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1321.396133] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76606716-b05a-4fed-93a8-5dd630ec46c0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1321.418660] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1321.418660] env[65788]: value = "task-4663442" [ 1321.418660] env[65788]: _type = "Task" [ 1321.418660] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1321.431156] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663442, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1321.912940] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "interface-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1321.913237] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.928862] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663442, 'name': ReconfigVM_Task, 'duration_secs': 0.214495} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1321.929190] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1322.416480] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1322.416839] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1322.417632] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417dad16-7756-4fc4-b589-a9b5958d5625 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.439641] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1322.439641] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1322.439869] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1322.440663] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1322.440663] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1322.441053] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1322.441314] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1322.441500] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1322.441704] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1322.441884] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1322.442076] env[65788]: DEBUG nova.virt.hardware [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1322.447424] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1322.448246] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a696a453-b9ea-4ad0-8499-4a9b14874471 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.462577] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e165bd89-27b6-46ea-ba3c-b035c2987347 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.484608] env[65788]: WARNING neutronclient.v2_0.client [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1322.490132] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Reconfiguring VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1322.491604] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5191aa6-b24b-4da6-ae11-2dc0e33dbe95 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.504684] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1322.504684] env[65788]: value = "task-4663443" [ 1322.504684] env[65788]: _type = "Task" [ 1322.504684] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.511591] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1322.511591] env[65788]: value = "task-4663444" [ 1322.511591] env[65788]: _type = "Task" [ 1322.511591] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.514782] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663443, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.522995] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.015552] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663443, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.024017] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.516648] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663443, 'name': ReconfigVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.525039] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.016835] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663443, 'name': ReconfigVM_Task, 'duration_secs': 1.174853} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.020335] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1324.021290] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3966d769-ee32-4b78-8e3a-bb26f36f41e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.030591] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.049405] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e/475291c8-a16d-4980-9ef4-7ed6a4dc382e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1324.050520] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-820eab66-f6eb-4dd0-b28f-db6ec22d1390 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.069289] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1324.069289] env[65788]: value = "task-4663445" [ 1324.069289] env[65788]: _type = "Task" [ 1324.069289] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.078179] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663445, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.526895] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.579686] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663445, 'name': ReconfigVM_Task, 'duration_secs': 0.281392} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1324.579936] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e/475291c8-a16d-4980-9ef4-7ed6a4dc382e.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1324.580219] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1325.027369] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.087891] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5afd4a83-85df-4a17-ae6b-ab53964a1541 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.118836] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b02e971-55b3-4576-8ee6-499a5b188d12 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.144720] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1325.529192] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.651490] env[65788]: WARNING neutronclient.v2_0.client [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1325.685275] env[65788]: DEBUG nova.network.neutron [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Port e7eece2d-f24b-4fcf-b1d8-a232af2c0021 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1326.029969] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.530854] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.708406] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1326.708406] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1326.708676] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1327.031478] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.530322] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 14%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.714844] env[65788]: WARNING neutronclient.v2_0.client [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1327.746277] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1327.746611] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1327.746859] env[65788]: DEBUG nova.network.neutron [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1328.030988] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task} progress is 18%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.250269] env[65788]: WARNING neutronclient.v2_0.client [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.250994] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.251369] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.367274] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.367680] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.435289] env[65788]: WARNING neutronclient.v2_0.client [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.436009] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.436375] env[65788]: WARNING openstack [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.517524] env[65788]: DEBUG nova.network.neutron [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1328.531970] env[65788]: DEBUG oslo_vmware.api [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663444, 'name': ReconfigVM_Task, 'duration_secs': 5.790101} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.532273] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1328.532476] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Reconfigured VM to detach interface {{(pid=65788) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1328.532884] env[65788]: WARNING neutronclient.v2_0.client [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.533207] env[65788]: WARNING neutronclient.v2_0.client [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1328.533814] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1328.534164] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1328.568380] env[65788]: WARNING neutronclient.v2_0.client [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1329.019993] env[65788]: DEBUG oslo_concurrency.lockutils [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1329.548475] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e820f00-e824-419e-9c33-d67c35fbcd67 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.569164] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96e0f68-20d8-4cf8-8d8b-a5bad165ec95 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.576486] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1329.826681] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1329.826880] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquired lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1329.827139] env[65788]: DEBUG nova.network.neutron [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1330.082336] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1330.082710] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8387d03d-f6d4-45f7-ad86-314f3230a2d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.090789] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1330.090789] env[65788]: value = "task-4663446" [ 1330.090789] env[65788]: _type = "Task" [ 1330.090789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.099831] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663446, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.318744] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1330.319061] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1330.319280] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1330.319465] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1330.319752] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1330.321914] env[65788]: INFO nova.compute.manager [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Terminating instance [ 1330.329903] env[65788]: WARNING neutronclient.v2_0.client [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1330.330575] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1330.330956] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1330.458091] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1330.458500] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1330.518673] env[65788]: WARNING neutronclient.v2_0.client [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1330.519343] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1330.519707] env[65788]: WARNING openstack [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1330.597675] env[65788]: INFO nova.network.neutron [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Port e0ea25a5-c2ed-405d-b6bf-bc16e70fd558 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1330.598072] env[65788]: DEBUG nova.network.neutron [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [{"id": "b6e00388-271a-41b9-bb29-4463d4a692ea", "address": "fa:16:3e:d0:9f:17", "network": {"id": "b03fb885-c6ac-4708-9401-d5b501cae296", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-477071438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.191", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "676963c25a3741998b10790689242dbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e00388-27", "ovs_interfaceid": "b6e00388-271a-41b9-bb29-4463d4a692ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1330.602464] env[65788]: DEBUG oslo_vmware.api [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663446, 'name': PowerOnVM_Task, 'duration_secs': 0.386098} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.603397] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1330.603397] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-0d8a7d57-0e12-4815-8702-e04045b7d106 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance '475291c8-a16d-4980-9ef4-7ed6a4dc382e' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1330.825534] env[65788]: DEBUG nova.compute.manager [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1330.825729] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1330.826619] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc82a66-e810-4544-bc62-42de60be68ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.834643] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1330.834876] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6de5f05f-9a5c-4dfb-9160-c8b473b121af {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.840871] env[65788]: DEBUG oslo_vmware.api [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1330.840871] env[65788]: value = "task-4663447" [ 1330.840871] env[65788]: _type = "Task" [ 1330.840871] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.849067] env[65788]: DEBUG oslo_vmware.api [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.103843] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Releasing lock "refresh_cache-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1331.351328] env[65788]: DEBUG oslo_vmware.api [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663447, 'name': PowerOffVM_Task, 'duration_secs': 0.182677} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1331.351677] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1331.351928] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1331.352278] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-642ae416-7dc9-42e3-b01b-42129b30a493 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.420687] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1331.420887] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1331.421082] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleting the datastore file [datastore2] cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1331.421370] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-02b581ce-de61-4209-a5fa-10ac95c4fb08 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.430187] env[65788]: DEBUG oslo_vmware.api [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1331.430187] env[65788]: value = "task-4663449" [ 1331.430187] env[65788]: _type = "Task" [ 1331.430187] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1331.439255] env[65788]: DEBUG oslo_vmware.api [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663449, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.607829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-d3721bf7-9ab7-416f-93ee-ccc90a8a57d4 tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "interface-cc9be8c2-a6a8-4b33-9b15-e9ee58e43304-e0ea25a5-c2ed-405d-b6bf-bc16e70fd558" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.694s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1332.840581] env[65788]: DEBUG oslo_vmware.api [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663449, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148084} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1332.841764] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1332.842122] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1332.842390] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1332.842644] env[65788]: INFO nova.compute.manager [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Took 2.02 seconds to destroy the instance on the hypervisor. [ 1332.842989] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1332.843796] env[65788]: DEBUG nova.compute.manager [-] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1332.843945] env[65788]: DEBUG nova.network.neutron [-] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1332.844456] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1332.845226] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1332.845524] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1332.883906] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.334122] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1333.334414] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1333.334606] env[65788]: DEBUG nova.compute.manager [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Going to confirm migration 9 {{(pid=65788) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5290}} [ 1333.635417] env[65788]: DEBUG nova.compute.manager [req-410cfa7d-fbfd-4dfa-9773-065fc3ebe651 req-d97c7d99-2942-4837-8ddd-10651a7d56e3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Received event network-vif-deleted-b6e00388-271a-41b9-bb29-4463d4a692ea {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1333.635417] env[65788]: INFO nova.compute.manager [req-410cfa7d-fbfd-4dfa-9773-065fc3ebe651 req-d97c7d99-2942-4837-8ddd-10651a7d56e3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Neutron deleted interface b6e00388-271a-41b9-bb29-4463d4a692ea; detaching it from the instance and deleting it from the info cache [ 1333.635627] env[65788]: DEBUG nova.network.neutron [req-410cfa7d-fbfd-4dfa-9773-065fc3ebe651 req-d97c7d99-2942-4837-8ddd-10651a7d56e3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1333.841475] env[65788]: WARNING neutronclient.v2_0.client [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.879410] env[65788]: WARNING neutronclient.v2_0.client [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1333.879763] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1333.879916] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1333.880097] env[65788]: DEBUG nova.network.neutron [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1333.880278] env[65788]: DEBUG nova.objects.instance [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'info_cache' on Instance uuid 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1334.112525] env[65788]: DEBUG nova.network.neutron [None req-66dac885-5ef3-47c6-8e16-8e132b97e7a2 None None] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1334.139475] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6e60432-d90f-4299-a675-e4bae7f374b9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.150484] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e18ebe-d703-4733-9d75-4f6e0af713ba {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.179051] env[65788]: DEBUG nova.compute.manager [req-410cfa7d-fbfd-4dfa-9773-065fc3ebe651 req-d97c7d99-2942-4837-8ddd-10651a7d56e3 service nova] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Detach interface failed, port_id=b6e00388-271a-41b9-bb29-4463d4a692ea, reason: Instance cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1334.615420] env[65788]: INFO nova.compute.manager [None req-66dac885-5ef3-47c6-8e16-8e132b97e7a2 None None] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Took 1.77 seconds to deallocate network for instance. [ 1334.887063] env[65788]: WARNING neutronclient.v2_0.client [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1334.887844] env[65788]: WARNING openstack [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1334.888239] env[65788]: WARNING openstack [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1335.016259] env[65788]: WARNING openstack [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1335.016878] env[65788]: WARNING openstack [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1335.080689] env[65788]: WARNING neutronclient.v2_0.client [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1335.081410] env[65788]: WARNING openstack [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1335.081798] env[65788]: WARNING openstack [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1335.121674] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1335.121797] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1335.122033] env[65788]: DEBUG nova.objects.instance [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'resources' on Instance uuid cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1335.161163] env[65788]: DEBUG nova.network.neutron [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [{"id": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "address": "fa:16:3e:59:c5:3c", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7eece2d-f2", "ovs_interfaceid": "e7eece2d-f24b-4fcf-b1d8-a232af2c0021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1335.664108] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-475291c8-a16d-4980-9ef4-7ed6a4dc382e" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1335.664374] env[65788]: DEBUG nova.objects.instance [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'migration_context' on Instance uuid 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1335.695187] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0477e308-44bb-4417-a57e-61e201568189 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.703596] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ad476d-9675-4ea2-b48d-5ab480110a01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.733313] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8c2058-9814-4494-bc3d-6f03d0d8d41f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.741583] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b753f791-dc1e-4da6-9048-56279eb56436 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.756218] env[65788]: DEBUG nova.compute.provider_tree [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1336.167195] env[65788]: DEBUG nova.objects.base [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Object Instance<475291c8-a16d-4980-9ef4-7ed6a4dc382e> lazy-loaded attributes: info_cache,migration_context {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1336.170075] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a19e60-31c1-4730-8d24-f8b6d538f1e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.188916] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a489e688-a9e7-4833-b3b7-f8e08e45b85c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.194996] env[65788]: DEBUG oslo_vmware.api [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1336.194996] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b79aa-1a10-c9bc-011e-d39f6a0ff584" [ 1336.194996] env[65788]: _type = "Task" [ 1336.194996] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1336.203208] env[65788]: DEBUG oslo_vmware.api [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b79aa-1a10-c9bc-011e-d39f6a0ff584, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.259893] env[65788]: DEBUG nova.scheduler.client.report [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1336.705971] env[65788]: DEBUG oslo_vmware.api [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524b79aa-1a10-c9bc-011e-d39f6a0ff584, 'name': SearchDatastore_Task, 'duration_secs': 0.008732} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.706282] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1336.764574] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.643s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1336.767057] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.061s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1336.791826] env[65788]: INFO nova.scheduler.client.report [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted allocations for instance cc9be8c2-a6a8-4b33-9b15-e9ee58e43304 [ 1337.301712] env[65788]: DEBUG oslo_concurrency.lockutils [None req-bdc9ae5d-5ee8-4663-a78c-03d52a68521e tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "cc9be8c2-a6a8-4b33-9b15-e9ee58e43304" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.983s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1337.324310] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3cae9d-7e72-4b46-bc5b-d19756a15761 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.333304] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431dc579-3350-4d95-9bf8-d367baebaa01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.365837] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024eaef3-0a83-48dc-826f-bae54bd965e1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.374092] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ae8574-2b07-4980-a11e-cc4c4f1c9718 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.388291] env[65788]: DEBUG nova.compute.provider_tree [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1337.891171] env[65788]: DEBUG nova.scheduler.client.report [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1338.086553] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.086835] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.087086] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1338.087244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1338.087410] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1338.089560] env[65788]: INFO nova.compute.manager [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Terminating instance [ 1338.593676] env[65788]: DEBUG nova.compute.manager [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1338.594094] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1338.594841] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe853a1-754b-4bef-ba0f-8fc886e497ed {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.602738] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1338.602979] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d916dfbc-345c-46ab-a933-051a7395ddac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.609665] env[65788]: DEBUG oslo_vmware.api [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1338.609665] env[65788]: value = "task-4663450" [ 1338.609665] env[65788]: _type = "Task" [ 1338.609665] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.618037] env[65788]: DEBUG oslo_vmware.api [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.901740] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.135s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1339.120455] env[65788]: DEBUG oslo_vmware.api [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663450, 'name': PowerOffVM_Task, 'duration_secs': 0.178306} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.120752] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1339.120897] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1339.121189] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05811901-6454-4726-bc2e-0f8549a35b22 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.200940] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1339.201158] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1339.201304] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleting the datastore file [datastore2] 4abbb248-12f6-4612-b6ea-01d1ef2104d7 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1339.201577] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d2f795f-c49e-4488-85c1-b172de7a9a7e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.208794] env[65788]: DEBUG oslo_vmware.api [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for the task: (returnval){ [ 1339.208794] env[65788]: value = "task-4663452" [ 1339.208794] env[65788]: _type = "Task" [ 1339.208794] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1339.216790] env[65788]: DEBUG oslo_vmware.api [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1339.463954] env[65788]: INFO nova.scheduler.client.report [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted allocation for migration ba8aa17b-96d2-40b6-8a39-4e67c1e48856 [ 1339.720456] env[65788]: DEBUG oslo_vmware.api [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Task: {'id': task-4663452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150907} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1339.720825] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1339.720923] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1339.721124] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1339.721317] env[65788]: INFO nova.compute.manager [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1339.721584] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1339.721802] env[65788]: DEBUG nova.compute.manager [-] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1339.721894] env[65788]: DEBUG nova.network.neutron [-] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1339.722177] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1339.722732] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1339.722983] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1339.817434] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1339.970535] env[65788]: DEBUG oslo_concurrency.lockutils [None req-fa8ef53a-7c08-48e3-95b4-9e88adafb18c tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.636s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1340.035040] env[65788]: DEBUG nova.compute.manager [req-88ebc76d-900d-4c6c-b574-c7ea2b102b0e req-49f510f9-9d9d-41b2-a815-0caef4a7ae95 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Received event network-vif-deleted-0c1d654b-57fe-46a4-a1f9-63029c094c97 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1340.035254] env[65788]: INFO nova.compute.manager [req-88ebc76d-900d-4c6c-b574-c7ea2b102b0e req-49f510f9-9d9d-41b2-a815-0caef4a7ae95 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Neutron deleted interface 0c1d654b-57fe-46a4-a1f9-63029c094c97; detaching it from the instance and deleting it from the info cache [ 1340.035490] env[65788]: DEBUG nova.network.neutron [req-88ebc76d-900d-4c6c-b574-c7ea2b102b0e req-49f510f9-9d9d-41b2-a815-0caef4a7ae95 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1340.234829] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1340.235076] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1340.235292] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1340.235539] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1340.235715] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1340.238209] env[65788]: INFO nova.compute.manager [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Terminating instance [ 1340.516271] env[65788]: DEBUG nova.network.neutron [-] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1340.538227] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed734506-76d1-46a0-868e-8555e5da63ad {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.548436] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524f7c19-b457-448d-a219-ec8e0e94d3b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.573994] env[65788]: DEBUG nova.compute.manager [req-88ebc76d-900d-4c6c-b574-c7ea2b102b0e req-49f510f9-9d9d-41b2-a815-0caef4a7ae95 service nova] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Detach interface failed, port_id=0c1d654b-57fe-46a4-a1f9-63029c094c97, reason: Instance 4abbb248-12f6-4612-b6ea-01d1ef2104d7 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1340.741686] env[65788]: DEBUG nova.compute.manager [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1340.742130] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1340.742817] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9792b5cb-86bf-441b-8b2d-ae447bbc58c9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.751259] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1340.751477] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1551f77e-618e-4c22-af78-260de63e52d1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.758389] env[65788]: DEBUG oslo_vmware.api [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1340.758389] env[65788]: value = "task-4663453" [ 1340.758389] env[65788]: _type = "Task" [ 1340.758389] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.018816] env[65788]: INFO nova.compute.manager [-] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Took 1.30 seconds to deallocate network for instance. [ 1341.270049] env[65788]: DEBUG oslo_vmware.api [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663453, 'name': PowerOffVM_Task, 'duration_secs': 0.260567} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.270254] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1341.270417] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1341.270675] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17174218-68bb-4f46-aa32-2f1fc93aad36 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.355636] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1341.355969] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1341.356171] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleting the datastore file [datastore2] 475291c8-a16d-4980-9ef4-7ed6a4dc382e {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1341.356510] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc942c42-b106-4653-aa35-fb52b02e5b2b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.364844] env[65788]: DEBUG oslo_vmware.api [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1341.364844] env[65788]: value = "task-4663455" [ 1341.364844] env[65788]: _type = "Task" [ 1341.364844] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.373684] env[65788]: DEBUG oslo_vmware.api [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663455, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.526362] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1341.526634] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1341.526864] env[65788]: DEBUG nova.objects.instance [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lazy-loading 'resources' on Instance uuid 4abbb248-12f6-4612-b6ea-01d1ef2104d7 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1341.875606] env[65788]: DEBUG oslo_vmware.api [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663455, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136519} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1341.876120] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1341.876120] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1341.876303] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1341.876381] env[65788]: INFO nova.compute.manager [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1341.876659] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1341.876861] env[65788]: DEBUG nova.compute.manager [-] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1341.876957] env[65788]: DEBUG nova.network.neutron [-] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1341.877285] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1341.877813] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1341.878080] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1341.918524] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1342.074402] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d846663e-745d-499e-9261-35a557258ae2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.082873] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a197c11-6579-4c32-a50a-450bae7345e9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.114135] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36efc223-1913-46a6-a6d5-476429b622d0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.122289] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a159ff79-a9ca-450d-bdb0-e0d00bd210bd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.136363] env[65788]: DEBUG nova.compute.provider_tree [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1342.639427] env[65788]: DEBUG nova.scheduler.client.report [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1342.829691] env[65788]: DEBUG nova.compute.manager [req-4bb0dfd7-c1ee-451d-8d5e-a5c8e75832fd req-fbededf5-e371-41a7-9c15-9a8234f8454b service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Received event network-vif-deleted-e7eece2d-f24b-4fcf-b1d8-a232af2c0021 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1342.829765] env[65788]: INFO nova.compute.manager [req-4bb0dfd7-c1ee-451d-8d5e-a5c8e75832fd req-fbededf5-e371-41a7-9c15-9a8234f8454b service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Neutron deleted interface e7eece2d-f24b-4fcf-b1d8-a232af2c0021; detaching it from the instance and deleting it from the info cache [ 1342.829906] env[65788]: DEBUG nova.network.neutron [req-4bb0dfd7-c1ee-451d-8d5e-a5c8e75832fd req-fbededf5-e371-41a7-9c15-9a8234f8454b service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1343.144524] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1343.165549] env[65788]: INFO nova.scheduler.client.report [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Deleted allocations for instance 4abbb248-12f6-4612-b6ea-01d1ef2104d7 [ 1343.303362] env[65788]: DEBUG nova.network.neutron [-] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1343.332470] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0725e84f-3fe2-4486-a194-03bbd69d69e6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.343087] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04d88c8-d34d-423d-8f75-57838a381b17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.368284] env[65788]: DEBUG nova.compute.manager [req-4bb0dfd7-c1ee-451d-8d5e-a5c8e75832fd req-fbededf5-e371-41a7-9c15-9a8234f8454b service nova] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Detach interface failed, port_id=e7eece2d-f24b-4fcf-b1d8-a232af2c0021, reason: Instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1343.672957] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4d09802d-5705-4c43-a25f-9f1170a2ffcc tempest-AttachInterfacesTestJSON-1497773155 tempest-AttachInterfacesTestJSON-1497773155-project-member] Lock "4abbb248-12f6-4612-b6ea-01d1ef2104d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.586s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1343.806398] env[65788]: INFO nova.compute.manager [-] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Took 1.93 seconds to deallocate network for instance. [ 1344.312345] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1344.312693] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1344.312834] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1344.340013] env[65788]: INFO nova.scheduler.client.report [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted allocations for instance 475291c8-a16d-4980-9ef4-7ed6a4dc382e [ 1344.847858] env[65788]: DEBUG oslo_concurrency.lockutils [None req-414e1739-f570-44f2-a62f-85f80e37772f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "475291c8-a16d-4980-9ef4-7ed6a4dc382e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.613s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1346.355574] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1346.355943] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1346.858617] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1347.383536] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1347.383859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1347.385435] env[65788]: INFO nova.compute.claims [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1348.447722] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aac7e54-bbec-4cba-960a-dd110e7f25ab {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.455770] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9117459-d11a-438e-ade2-7ced048d9bf7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.487056] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1310b0b1-0ce5-4622-980e-e68b92bd270e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.494983] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2258775-367a-4161-81c1-58d8fabcb7fd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.509074] env[65788]: DEBUG nova.compute.provider_tree [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1349.012201] env[65788]: DEBUG nova.scheduler.client.report [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1349.518569] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1349.519189] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1350.025049] env[65788]: DEBUG nova.compute.utils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1350.026873] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1350.027092] env[65788]: DEBUG nova.network.neutron [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1350.027421] env[65788]: WARNING neutronclient.v2_0.client [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1350.027722] env[65788]: WARNING neutronclient.v2_0.client [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1350.028354] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1350.028687] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1350.089145] env[65788]: DEBUG nova.policy [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02eddcb388a431abdff72ad3995dda1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69c8fb9f477e45068cd7c1e4defb64a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1350.374538] env[65788]: DEBUG nova.network.neutron [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Successfully created port: 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1350.537427] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1351.547751] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1351.573246] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1351.573539] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1351.573703] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1351.573912] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1351.574110] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1351.574300] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1351.574520] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1351.574675] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1351.574838] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1351.574994] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1351.575174] env[65788]: DEBUG nova.virt.hardware [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1351.576131] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0597ff4-aece-4398-85c1-3b3f4d03114c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.584661] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7069de-cbd1-4e2d-8fa0-3dfd2edac3be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.793034] env[65788]: DEBUG nova.compute.manager [req-e8d1a177-f5bf-4fcf-8948-7160007b8d05 req-5bd716df-9b7d-419d-947c-e4b7c3bea852 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Received event network-vif-plugged-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1351.793280] env[65788]: DEBUG oslo_concurrency.lockutils [req-e8d1a177-f5bf-4fcf-8948-7160007b8d05 req-5bd716df-9b7d-419d-947c-e4b7c3bea852 service nova] Acquiring lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1351.793555] env[65788]: DEBUG oslo_concurrency.lockutils [req-e8d1a177-f5bf-4fcf-8948-7160007b8d05 req-5bd716df-9b7d-419d-947c-e4b7c3bea852 service nova] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1351.793726] env[65788]: DEBUG oslo_concurrency.lockutils [req-e8d1a177-f5bf-4fcf-8948-7160007b8d05 req-5bd716df-9b7d-419d-947c-e4b7c3bea852 service nova] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1351.793899] env[65788]: DEBUG nova.compute.manager [req-e8d1a177-f5bf-4fcf-8948-7160007b8d05 req-5bd716df-9b7d-419d-947c-e4b7c3bea852 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] No waiting events found dispatching network-vif-plugged-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1351.794073] env[65788]: WARNING nova.compute.manager [req-e8d1a177-f5bf-4fcf-8948-7160007b8d05 req-5bd716df-9b7d-419d-947c-e4b7c3bea852 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Received unexpected event network-vif-plugged-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 for instance with vm_state building and task_state spawning. [ 1351.875259] env[65788]: DEBUG nova.network.neutron [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Successfully updated port: 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1352.378178] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1352.378318] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1352.378461] env[65788]: DEBUG nova.network.neutron [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1352.881316] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1352.881838] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1352.917228] env[65788]: DEBUG nova.network.neutron [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1352.935793] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1352.936192] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1352.995314] env[65788]: WARNING neutronclient.v2_0.client [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1352.996032] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1352.996389] env[65788]: WARNING openstack [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1353.075484] env[65788]: DEBUG nova.network.neutron [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1353.579394] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1353.579710] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Instance network_info: |[{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1353.580220] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:3e:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1353.587815] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1353.588045] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1353.588283] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d95dc949-3668-4e95-9dd2-3fb463740144 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.611237] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1353.611237] env[65788]: value = "task-4663456" [ 1353.611237] env[65788]: _type = "Task" [ 1353.611237] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1353.621802] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663456, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.822897] env[65788]: DEBUG nova.compute.manager [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Received event network-changed-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1353.823128] env[65788]: DEBUG nova.compute.manager [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Refreshing instance network info cache due to event network-changed-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1353.823349] env[65788]: DEBUG oslo_concurrency.lockutils [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1353.823494] env[65788]: DEBUG oslo_concurrency.lockutils [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1353.823657] env[65788]: DEBUG nova.network.neutron [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Refreshing network info cache for port 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1354.122551] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663456, 'name': CreateVM_Task, 'duration_secs': 0.325988} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.122777] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1354.123288] env[65788]: WARNING neutronclient.v2_0.client [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1354.123680] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1354.123872] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1354.124196] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1354.124453] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8aa23e72-a76d-46fc-be63-8d2c34a01ea5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.130133] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1354.130133] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52209fd5-af51-fdcf-c0cf-f5ee6018d4d5" [ 1354.130133] env[65788]: _type = "Task" [ 1354.130133] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.138409] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52209fd5-af51-fdcf-c0cf-f5ee6018d4d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.327529] env[65788]: WARNING neutronclient.v2_0.client [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1354.329064] env[65788]: WARNING openstack [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1354.329064] env[65788]: WARNING openstack [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1354.415681] env[65788]: WARNING openstack [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1354.416102] env[65788]: WARNING openstack [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1354.475259] env[65788]: WARNING neutronclient.v2_0.client [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1354.476044] env[65788]: WARNING openstack [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1354.476394] env[65788]: WARNING openstack [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1354.554560] env[65788]: DEBUG nova.network.neutron [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updated VIF entry in instance network info cache for port 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1354.554934] env[65788]: DEBUG nova.network.neutron [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1354.640755] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52209fd5-af51-fdcf-c0cf-f5ee6018d4d5, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.641073] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1354.641312] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1354.641543] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1354.641699] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1354.641911] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1354.642200] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02fba736-827d-498b-9cdb-b9f499beaec7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.651022] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1354.651195] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1354.651877] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ab8fe5e-9597-435f-8ce8-0d183080a6b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.657312] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1354.657312] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f89a5e-d4f6-c47d-c042-d76e36a2e9a3" [ 1354.657312] env[65788]: _type = "Task" [ 1354.657312] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.665416] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f89a5e-d4f6-c47d-c042-d76e36a2e9a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.058077] env[65788]: DEBUG oslo_concurrency.lockutils [req-16474c94-5f54-4f54-9b86-14004d0d68d0 req-f1705f2a-4d16-49b1-8713-2cf967ae47e1 service nova] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1355.168758] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52f89a5e-d4f6-c47d-c042-d76e36a2e9a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008587} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.169495] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0861ae1-7400-4a24-8538-5bfee406f695 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.174560] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1355.174560] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529f1e83-1e19-a020-e0b6-b5d49f6f8618" [ 1355.174560] env[65788]: _type = "Task" [ 1355.174560] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.182699] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529f1e83-1e19-a020-e0b6-b5d49f6f8618, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.686039] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]529f1e83-1e19-a020-e0b6-b5d49f6f8618, 'name': SearchDatastore_Task, 'duration_secs': 0.009758} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.686244] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1355.686359] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1355.686642] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-371aa62f-276f-4dff-8295-e2366229e38a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.694225] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1355.694225] env[65788]: value = "task-4663457" [ 1355.694225] env[65788]: _type = "Task" [ 1355.694225] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1355.702694] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663457, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.206324] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663457, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47954} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.206751] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1356.206842] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1356.207051] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-afbbcfbe-1711-4907-bdfc-c6f928665ed4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.213563] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1356.213563] env[65788]: value = "task-4663458" [ 1356.213563] env[65788]: _type = "Task" [ 1356.213563] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.222927] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1356.723448] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059391} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1356.723671] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1356.724569] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ded5449-701f-4ace-9baf-5175a928dcc1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.746060] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1356.746345] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-228f5329-ecc9-4d58-bdf6-e4fd51336bd6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.766409] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1356.766409] env[65788]: value = "task-4663459" [ 1356.766409] env[65788]: _type = "Task" [ 1356.766409] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.774716] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663459, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.276880] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663459, 'name': ReconfigVM_Task, 'duration_secs': 0.28124} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.277308] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1357.277806] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30ad8751-96ca-4912-a884-9fb61ae489d5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.284604] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1357.284604] env[65788]: value = "task-4663460" [ 1357.284604] env[65788]: _type = "Task" [ 1357.284604] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.292495] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663460, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.794455] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663460, 'name': Rename_Task, 'duration_secs': 0.163775} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.794671] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1357.794917] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f53e645-5329-4fb0-bd8f-149ef56c1133 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.801420] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1357.801420] env[65788]: value = "task-4663461" [ 1357.801420] env[65788]: _type = "Task" [ 1357.801420] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.809296] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663461, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.313698] env[65788]: DEBUG oslo_vmware.api [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663461, 'name': PowerOnVM_Task, 'duration_secs': 0.442988} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.314117] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1358.314218] env[65788]: INFO nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Took 6.77 seconds to spawn the instance on the hypervisor. [ 1358.314372] env[65788]: DEBUG nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1358.315192] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202da2e2-ec01-4e7c-931b-cd5e33269e32 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.834370] env[65788]: INFO nova.compute.manager [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Took 11.47 seconds to build instance. [ 1359.336716] env[65788]: DEBUG oslo_concurrency.lockutils [None req-1be42c9f-7d9f-475f-a665-3aa9982e73c3 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.981s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1359.398271] env[65788]: DEBUG nova.compute.manager [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Received event network-changed-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1359.398426] env[65788]: DEBUG nova.compute.manager [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Refreshing instance network info cache due to event network-changed-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1359.398657] env[65788]: DEBUG oslo_concurrency.lockutils [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.398849] env[65788]: DEBUG oslo_concurrency.lockutils [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1359.398961] env[65788]: DEBUG nova.network.neutron [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Refreshing network info cache for port 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1359.901903] env[65788]: WARNING neutronclient.v2_0.client [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1359.903834] env[65788]: WARNING openstack [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1359.904233] env[65788]: WARNING openstack [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1360.013034] env[65788]: WARNING openstack [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1360.013438] env[65788]: WARNING openstack [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1360.075016] env[65788]: WARNING neutronclient.v2_0.client [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1360.075729] env[65788]: WARNING openstack [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1360.076079] env[65788]: WARNING openstack [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1360.156101] env[65788]: DEBUG nova.network.neutron [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updated VIF entry in instance network info cache for port 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1360.156479] env[65788]: DEBUG nova.network.neutron [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1360.659604] env[65788]: DEBUG oslo_concurrency.lockutils [req-748a9133-ab0f-4542-a833-9c05567b3561 req-23f0361b-5b34-466f-a6a2-81f7694ea098 service nova] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1374.384658] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.384658] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.384658] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.385248] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1375.638567] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1376.633214] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1377.138547] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1377.138964] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1379.638126] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.638531] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1380.141556] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1380.141825] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1380.141998] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1380.142171] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1380.143129] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db64c5c4-d749-4206-930c-4e4c45402df3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.152043] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccffe653-8e9c-4214-abbc-1e86630ccf56 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.166973] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790c0134-a3be-400e-80bc-80c06006b159 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.173931] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc5db39-b046-42da-8ce0-6ac9556823b2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1380.203407] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179831MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1380.203624] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1380.203769] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1381.229835] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1381.230146] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1381.230209] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_active': '1', 'num_task_None': '1', 'num_os_type_None': '1', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '1', 'io_workload': '0'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1381.258220] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85a7a6d-23aa-4d1c-aad9-0881aff63bde {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.266444] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4426f81a-7c8f-4779-a14d-831b7b650d01 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.296526] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69adc404-f441-4ef6-972e-aec98fbb471e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.304596] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3721f62-40ac-41b9-9564-44b177088021 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1381.318373] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1381.821876] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1382.327505] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1382.327921] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1397.755847] env[65788]: DEBUG nova.compute.manager [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Stashing vm_state: active {{(pid=65788) _prep_resize /opt/stack/nova/nova/compute/manager.py:6176}} [ 1398.274963] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1398.275276] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1398.780708] env[65788]: INFO nova.compute.claims [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1399.287372] env[65788]: INFO nova.compute.resource_tracker [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating resource usage from migration 90369a7d-d8c6-43de-b81b-dbe68cc77863 [ 1399.329637] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc17fc8-6487-45eb-915f-22829d07d0e8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.337804] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6e2c16-5411-4555-bdf2-7b1ab2eb2d44 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.369810] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c51e918-435f-4bba-badc-d8c1a1e6a13a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.377914] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87dce73-cdc6-498e-92b4-238363666028 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.391714] env[65788]: DEBUG nova.compute.provider_tree [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1399.895131] env[65788]: DEBUG nova.scheduler.client.report [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1400.400636] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.125s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1400.400867] env[65788]: INFO nova.compute.manager [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Migrating [ 1400.916098] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1400.916516] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1400.916881] env[65788]: DEBUG nova.network.neutron [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1401.420403] env[65788]: WARNING neutronclient.v2_0.client [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1401.421128] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1401.421480] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1401.543470] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1401.543921] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1401.626502] env[65788]: WARNING neutronclient.v2_0.client [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1401.627270] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1401.627677] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1401.710715] env[65788]: DEBUG nova.network.neutron [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1402.213957] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1403.730743] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09299d76-871a-43b3-919c-fb720771547c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.751942] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 0 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1404.258684] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1404.258996] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-027c4630-6a08-4607-9296-408f95144aef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.267304] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1404.267304] env[65788]: value = "task-4663462" [ 1404.267304] env[65788]: _type = "Task" [ 1404.267304] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.275558] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1404.778516] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663462, 'name': PowerOffVM_Task, 'duration_secs': 0.204304} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1404.780227] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1404.780227] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 17 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1405.286846] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:49Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1405.287155] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1405.287325] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1405.287505] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1405.287647] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1405.287791] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1405.287988] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1405.288161] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1405.288322] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1405.288479] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1405.288684] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1405.293706] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8754f32-19a1-464e-8650-9a2b7b1596fa {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.310654] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1405.310654] env[65788]: value = "task-4663463" [ 1405.310654] env[65788]: _type = "Task" [ 1405.310654] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.319518] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663463, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.821374] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663463, 'name': ReconfigVM_Task, 'duration_secs': 0.160616} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.821787] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 33 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1406.328681] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1406.328981] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1406.329074] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1406.329258] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1406.329407] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1406.329549] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1406.329753] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1406.329912] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1406.330092] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1406.330256] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1406.330428] env[65788]: DEBUG nova.virt.hardware [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1406.335608] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfiguring VM instance instance-0000007d to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1406.335940] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18280d7c-1cc2-47a5-9045-fb501aafa2e4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.356804] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1406.356804] env[65788]: value = "task-4663464" [ 1406.356804] env[65788]: _type = "Task" [ 1406.356804] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.365830] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663464, 'name': ReconfigVM_Task} progress is 6%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.867907] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663464, 'name': ReconfigVM_Task, 'duration_secs': 0.15927} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.868303] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfigured VM instance instance-0000007d to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1406.868966] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1029598f-5a1b-4b82-a0e3-317aecf477f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.893444] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1406.894196] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff836c6d-c112-44c2-89c3-89388b9e75ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1406.912713] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1406.912713] env[65788]: value = "task-4663465" [ 1406.912713] env[65788]: _type = "Task" [ 1406.912713] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1406.922430] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663465, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1407.422250] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663465, 'name': ReconfigVM_Task, 'duration_secs': 0.251667} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1407.422527] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1407.422792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 50 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1407.929424] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697dab02-f1c6-4e95-8bde-8d3ea15e8359 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.949210] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df298a2b-2cc0-4dc0-b681-f1e24aee531a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.967277] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 67 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1408.475070] env[65788]: WARNING neutronclient.v2_0.client [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1408.509348] env[65788]: DEBUG nova.network.neutron [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Port 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1409.533819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1409.533819] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1409.534238] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1410.540614] env[65788]: WARNING neutronclient.v2_0.client [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1410.582846] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1410.583045] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1410.583228] env[65788]: DEBUG nova.network.neutron [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1411.087734] env[65788]: WARNING neutronclient.v2_0.client [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1411.088514] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1411.088875] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1411.205209] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1411.205614] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1411.263400] env[65788]: WARNING neutronclient.v2_0.client [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1411.264115] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1411.264469] env[65788]: WARNING openstack [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1411.342996] env[65788]: DEBUG nova.network.neutron [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1411.846101] env[65788]: DEBUG oslo_concurrency.lockutils [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1412.369948] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e067172-7ed2-499f-adc2-01456b9f92d4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.389742] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497081b3-f360-4443-8d9f-dae5176eac80 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.396793] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 83 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1412.903240] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1412.903609] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85525644-3e85-4eb2-95cd-61ebb5b7ae8e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.911321] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1412.911321] env[65788]: value = "task-4663466" [ 1412.911321] env[65788]: _type = "Task" [ 1412.911321] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1412.920034] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663466, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1413.422349] env[65788]: DEBUG oslo_vmware.api [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663466, 'name': PowerOnVM_Task, 'duration_secs': 0.382795} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1413.422624] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1413.422806] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-897fc925-9d6a-4168-b5b7-9081c2644f2f tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance '2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0' progress to 100 {{(pid=65788) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1415.712085] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1415.745132] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1415.745512] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1415.779949] env[65788]: DEBUG nova.network.neutron [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Port 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 binding to destination host cpu-1 is already ACTIVE {{(pid=65788) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3284}} [ 1415.780238] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1415.780386] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1415.780546] env[65788]: DEBUG nova.network.neutron [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1416.283205] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1416.284379] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1416.284754] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1416.400712] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1416.401129] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1416.496477] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1416.497172] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1416.497528] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1416.579649] env[65788]: DEBUG nova.network.neutron [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1417.082798] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1417.586560] env[65788]: DEBUG nova.compute.manager [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=65788) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:925}} [ 1417.586811] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1417.587106] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1418.090595] env[65788]: DEBUG nova.objects.instance [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'migration_context' on Instance uuid 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1418.636147] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e89695b-4111-4ecf-bcef-25e16e71ec8b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.644244] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a735e41a-e813-49d5-a568-67fd7687635a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.674881] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df4df32-05f0-4831-9fc8-3b5452bbe165 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.682986] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b31ed5-d01b-4f0a-a402-27b35a8a090c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.697872] env[65788]: DEBUG nova.compute.provider_tree [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1419.201212] env[65788]: DEBUG nova.scheduler.client.report [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1420.213483] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.626s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1421.748921] env[65788]: INFO nova.compute.manager [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Swapping old allocation on dict_keys(['3a6417f0-b161-4086-8a0e-1cb139eae377']) held by migration 90369a7d-d8c6-43de-b81b-dbe68cc77863 for instance [ 1421.773471] env[65788]: DEBUG nova.scheduler.client.report [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Overwriting current allocation {'allocations': {'3a6417f0-b161-4086-8a0e-1cb139eae377': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 176}}, 'project_id': '69c8fb9f477e45068cd7c1e4defb64a0', 'user_id': 'e02eddcb388a431abdff72ad3995dda1', 'consumer_generation': 1} on consumer 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0 {{(pid=65788) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1421.837903] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1421.878392] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1421.878587] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1421.878768] env[65788]: DEBUG nova.network.neutron [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1422.382247] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1422.382888] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1422.383264] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1422.511362] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1422.511763] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1422.573657] env[65788]: WARNING neutronclient.v2_0.client [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1422.574340] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1422.574681] env[65788]: WARNING openstack [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1422.652543] env[65788]: DEBUG nova.network.neutron [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [{"id": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "address": "fa:16:3e:ee:3e:50", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70f44b4b-cb", "ovs_interfaceid": "70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1423.155312] env[65788]: DEBUG oslo_concurrency.lockutils [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1423.155776] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1423.156121] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-429093d3-4f27-410f-abdd-c271c74a29cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.164686] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1423.164686] env[65788]: value = "task-4663467" [ 1423.164686] env[65788]: _type = "Task" [ 1423.164686] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.173787] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1423.676103] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663467, 'name': PowerOffVM_Task, 'duration_secs': 0.20738} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1423.676385] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1423.677054] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1423.677289] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1423.677441] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1423.677621] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1423.677763] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1423.677904] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1423.678122] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1423.678280] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1423.678439] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1423.678592] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1423.678760] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1423.683819] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f9398fb-7ccf-46bc-92cf-421c396bc237 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1423.700214] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1423.700214] env[65788]: value = "task-4663468" [ 1423.700214] env[65788]: _type = "Task" [ 1423.700214] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1423.708772] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663468, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.210967] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663468, 'name': ReconfigVM_Task, 'duration_secs': 0.136098} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.211853] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547acac9-d0ec-4055-b40c-82e701cca560 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.231387] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1424.231645] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1424.231770] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1424.231954] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1424.232147] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1424.232307] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1424.232534] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1424.232713] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1424.232963] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1424.233175] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1424.233359] env[65788]: DEBUG nova.virt.hardware [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1424.234188] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f93796c-bcc9-4e02-b69f-2a8a2e0cc160 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.240799] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1424.240799] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52379434-3828-55ee-8d98-771ee0ab20f8" [ 1424.240799] env[65788]: _type = "Task" [ 1424.240799] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.249688] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52379434-3828-55ee-8d98-771ee0ab20f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1424.751450] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52379434-3828-55ee-8d98-771ee0ab20f8, 'name': SearchDatastore_Task, 'duration_secs': 0.007851} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1424.757032] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfiguring VM instance instance-0000007d to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1424.757347] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a485b427-e944-4b44-9922-ecab4e25f1cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.776877] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1424.776877] env[65788]: value = "task-4663469" [ 1424.776877] env[65788]: _type = "Task" [ 1424.776877] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1424.785647] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663469, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1425.288219] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663469, 'name': ReconfigVM_Task, 'duration_secs': 0.179392} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.288543] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfigured VM instance instance-0000007d to detach disk 2000 {{(pid=65788) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1425.289346] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67409f16-45c4-4df6-a704-6118432d6404 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.312189] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1425.312598] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e89d96fd-1fb7-4b7c-9f6a-afc27a6f9f2d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.331789] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1425.331789] env[65788]: value = "task-4663470" [ 1425.331789] env[65788]: _type = "Task" [ 1425.331789] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1425.340545] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663470, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1425.842669] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663470, 'name': ReconfigVM_Task, 'duration_secs': 0.272906} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1425.842971] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0/2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0.vmdk or device None with type thin {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1425.843848] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf357a2-2e2c-42e8-bd82-e82f874922e2 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.865675] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b32954d-65ca-43f2-ac57-c651092091f4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.884106] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2160a6ff-4865-4c5a-91c0-2e020866d7c5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.902671] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404f09ef-c178-44bd-9506-ad5a4d76cb2a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.910037] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1425.910305] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bc636454-1b9c-43e2-8da6-be70df249723 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1425.917206] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1425.917206] env[65788]: value = "task-4663471" [ 1425.917206] env[65788]: _type = "Task" [ 1425.917206] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1425.925677] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663471, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1426.428016] env[65788]: DEBUG oslo_vmware.api [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663471, 'name': PowerOnVM_Task, 'duration_secs': 0.444852} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1426.428494] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1427.440677] env[65788]: INFO nova.compute.manager [None req-e5c3d4d1-d85d-4afd-ac6f-a448441b2614 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance to original state: 'active' [ 1429.446943] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1429.447366] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1429.447516] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1429.447663] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1429.447832] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1429.451751] env[65788]: INFO nova.compute.manager [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Terminating instance [ 1429.956506] env[65788]: DEBUG nova.compute.manager [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1429.956792] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1429.958141] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69386509-c69e-4807-b953-7c9eeb0b76dd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.966076] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1429.966357] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f9f8c6c-5a29-44c5-b8d6-872d01c91d7e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.973385] env[65788]: DEBUG oslo_vmware.api [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1429.973385] env[65788]: value = "task-4663472" [ 1429.973385] env[65788]: _type = "Task" [ 1429.973385] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.982018] env[65788]: DEBUG oslo_vmware.api [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.484823] env[65788]: DEBUG oslo_vmware.api [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663472, 'name': PowerOffVM_Task, 'duration_secs': 0.197497} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.485238] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1430.485364] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1430.485577] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d99419d-a699-494b-a8c7-9c8703b6ac0c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.556019] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1430.556332] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1430.556507] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleting the datastore file [datastore2] 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1430.556793] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13794740-7117-49b2-b5d2-7e58ef2a53ca {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.563606] env[65788]: DEBUG oslo_vmware.api [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1430.563606] env[65788]: value = "task-4663474" [ 1430.563606] env[65788]: _type = "Task" [ 1430.563606] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.571723] env[65788]: DEBUG oslo_vmware.api [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663474, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.074394] env[65788]: DEBUG oslo_vmware.api [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663474, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144317} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.074646] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1431.074825] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1431.075023] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1431.075206] env[65788]: INFO nova.compute.manager [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1431.075483] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1431.075678] env[65788]: DEBUG nova.compute.manager [-] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1431.075807] env[65788]: DEBUG nova.network.neutron [-] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1431.076034] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1431.076593] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1431.076850] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1431.112083] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1431.574130] env[65788]: DEBUG nova.compute.manager [req-7c9cec80-3165-4edc-a4fb-b5468886cdff req-53ce74b4-be4e-474a-bb0d-56bddc190944 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Received event network-vif-deleted-70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1431.574397] env[65788]: INFO nova.compute.manager [req-7c9cec80-3165-4edc-a4fb-b5468886cdff req-53ce74b4-be4e-474a-bb0d-56bddc190944 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Neutron deleted interface 70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0; detaching it from the instance and deleting it from the info cache [ 1431.574608] env[65788]: DEBUG nova.network.neutron [req-7c9cec80-3165-4edc-a4fb-b5468886cdff req-53ce74b4-be4e-474a-bb0d-56bddc190944 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1431.638349] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1432.054878] env[65788]: DEBUG nova.network.neutron [-] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1432.077091] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7497267-e8bb-49c2-9b73-ed7efef073df {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.087712] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012f2eaa-5b9b-46bc-a3fd-dbdc8102feb0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.112714] env[65788]: DEBUG nova.compute.manager [req-7c9cec80-3165-4edc-a4fb-b5468886cdff req-53ce74b4-be4e-474a-bb0d-56bddc190944 service nova] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Detach interface failed, port_id=70f44b4b-cbc4-4d9a-bc86-ab4a45508ef0, reason: Instance 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0 could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1432.558859] env[65788]: INFO nova.compute.manager [-] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Took 1.48 seconds to deallocate network for instance. [ 1433.065895] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1433.066476] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1433.066674] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1433.089142] env[65788]: INFO nova.scheduler.client.report [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted allocations for instance 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0 [ 1433.139697] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1433.139906] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1433.596633] env[65788]: DEBUG oslo_concurrency.lockutils [None req-4a8c6e9e-2fd8-4325-9c17-3089702b7d5b tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.149s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1434.384783] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "1abcfa65-a4e9-452d-8be3-33724601be9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1434.385126] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1434.637652] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1434.637938] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1434.887627] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1435.415271] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1435.415645] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1435.417906] env[65788]: INFO nova.compute.claims [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1435.638713] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1436.553178] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92eb51c-b6c5-424c-9323-ce61de246b17 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.563515] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4c516a-a728-4617-bbba-f33941697794 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.602274] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5442ae-e795-4dc7-93f5-4d7d8582ab58 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.610477] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db28fe9-e987-4647-bc9d-cf991feaf581 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.624028] env[65788]: DEBUG nova.compute.provider_tree [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1437.127244] env[65788]: DEBUG nova.scheduler.client.report [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1437.631748] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.216s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1437.632296] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1438.138208] env[65788]: DEBUG nova.compute.utils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1438.139716] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Allocating IP information in the background. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2020}} [ 1438.139900] env[65788]: DEBUG nova.network.neutron [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] allocate_for_instance() {{(pid=65788) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1238}} [ 1438.140246] env[65788]: WARNING neutronclient.v2_0.client [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1438.140604] env[65788]: WARNING neutronclient.v2_0.client [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1438.141203] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1438.141548] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1438.196474] env[65788]: DEBUG nova.policy [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e02eddcb388a431abdff72ad3995dda1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69c8fb9f477e45068cd7c1e4defb64a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=65788) authorize /opt/stack/nova/nova/policy.py:192}} [ 1438.490341] env[65788]: DEBUG nova.network.neutron [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Successfully created port: bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 1438.637835] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1438.638242] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1438.649911] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1439.638055] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.659131] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1439.685998] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1439.686305] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1439.686466] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1439.686650] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1439.686793] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1439.686940] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1439.687160] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1439.687347] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1439.687521] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1439.687682] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1439.687849] env[65788]: DEBUG nova.virt.hardware [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1439.688748] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9b1bb0-4b96-4883-ae8f-de54b4e49716 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.697159] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b0b4f6-5314-4aa9-8fb4-136f89b72be0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1439.907260] env[65788]: DEBUG nova.compute.manager [req-9b0e84b2-1c7c-42db-94db-600accbe5051 req-366bf3b8-d152-440c-bbc0-bcd1a9737eee service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Received event network-vif-plugged-bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1439.907494] env[65788]: DEBUG oslo_concurrency.lockutils [req-9b0e84b2-1c7c-42db-94db-600accbe5051 req-366bf3b8-d152-440c-bbc0-bcd1a9737eee service nova] Acquiring lock "1abcfa65-a4e9-452d-8be3-33724601be9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1439.907697] env[65788]: DEBUG oslo_concurrency.lockutils [req-9b0e84b2-1c7c-42db-94db-600accbe5051 req-366bf3b8-d152-440c-bbc0-bcd1a9737eee service nova] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1439.907866] env[65788]: DEBUG oslo_concurrency.lockutils [req-9b0e84b2-1c7c-42db-94db-600accbe5051 req-366bf3b8-d152-440c-bbc0-bcd1a9737eee service nova] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1439.908073] env[65788]: DEBUG nova.compute.manager [req-9b0e84b2-1c7c-42db-94db-600accbe5051 req-366bf3b8-d152-440c-bbc0-bcd1a9737eee service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] No waiting events found dispatching network-vif-plugged-bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) pop_instance_event /opt/stack/nova/nova/compute/manager.py:345}} [ 1439.908205] env[65788]: WARNING nova.compute.manager [req-9b0e84b2-1c7c-42db-94db-600accbe5051 req-366bf3b8-d152-440c-bbc0-bcd1a9737eee service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Received unexpected event network-vif-plugged-bab8becc-f8fa-4a74-abbd-677bfd2508f1 for instance with vm_state building and task_state spawning. [ 1439.985661] env[65788]: DEBUG nova.network.neutron [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Successfully updated port: bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) _update_port /opt/stack/nova/nova/network/neutron.py:617}} [ 1440.141208] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1440.141427] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1440.141604] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1440.141761] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1440.142759] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a148c78f-b307-4a8e-99dc-a0277d21d146 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.151881] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad50838-2e5b-4304-92fe-050bc9195a6f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.166459] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16a1180-b352-45fb-a30c-f0303299b512 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.174095] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddb5c9d-5c14-43f2-8f18-b86f9359b1fe {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1440.204574] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180420MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1440.204734] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1440.204956] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1440.489520] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1440.489662] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1440.489901] env[65788]: DEBUG nova.network.neutron [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1440.993695] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1440.994129] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1441.038536] env[65788]: DEBUG nova.network.neutron [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1441.057846] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1441.058261] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1441.120720] env[65788]: WARNING neutronclient.v2_0.client [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1441.121437] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1441.121840] env[65788]: WARNING openstack [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1441.207129] env[65788]: DEBUG nova.network.neutron [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [{"id": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "address": "fa:16:3e:27:d3:e2", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbab8becc-f8", "ovs_interfaceid": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1441.228688] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance 1abcfa65-a4e9-452d-8be3-33724601be9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 1441.228886] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1441.229044] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=100GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '1', 'num_vm_building': '1', 'num_task_spawning': '1', 'num_os_type_None': '1', 'num_proj_69c8fb9f477e45068cd7c1e4defb64a0': '1', 'io_workload': '1'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1441.256519] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c8368c-64c4-49d3-9722-8101607bdc69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.264978] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db239fee-2dae-4492-918c-4d9228a6bd69 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.296422] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c143948-e9e5-490e-b9a3-a928c6f51f5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.305803] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245196d2-71ff-4ab7-9898-3df167520da8 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.320115] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1441.710371] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1441.710751] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Instance network_info: |[{"id": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "address": "fa:16:3e:27:d3:e2", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbab8becc-f8", "ovs_interfaceid": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2035}} [ 1441.711263] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:d3:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bab8becc-f8fa-4a74-abbd-677bfd2508f1', 'vif_model': 'vmxnet3'}] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1441.719100] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1441.719321] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1441.719556] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d37e0d1f-0ebe-4b7f-ab14-327a2fcac680 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1441.739953] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1441.739953] env[65788]: value = "task-4663475" [ 1441.739953] env[65788]: _type = "Task" [ 1441.739953] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1441.748537] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663475, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1441.823848] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1441.937489] env[65788]: DEBUG nova.compute.manager [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Received event network-changed-bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1441.937673] env[65788]: DEBUG nova.compute.manager [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Refreshing instance network info cache due to event network-changed-bab8becc-f8fa-4a74-abbd-677bfd2508f1. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1441.938139] env[65788]: DEBUG oslo_concurrency.lockutils [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Acquiring lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1441.938569] env[65788]: DEBUG oslo_concurrency.lockutils [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Acquired lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1441.938569] env[65788]: DEBUG nova.network.neutron [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Refreshing network info cache for port bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1442.250700] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663475, 'name': CreateVM_Task, 'duration_secs': 0.308603} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.251093] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1442.251451] env[65788]: WARNING neutronclient.v2_0.client [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1442.251800] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.251954] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1442.252294] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1442.252560] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7797afa6-b187-42d1-84ba-35549f933802 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.258482] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1442.258482] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528cb565-fe43-86fb-3e52-dca0bbcc6656" [ 1442.258482] env[65788]: _type = "Task" [ 1442.258482] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.267158] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528cb565-fe43-86fb-3e52-dca0bbcc6656, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1442.329157] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1442.329356] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1442.441530] env[65788]: WARNING neutronclient.v2_0.client [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1442.442264] env[65788]: WARNING openstack [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1442.442669] env[65788]: WARNING openstack [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1442.532446] env[65788]: WARNING openstack [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1442.532889] env[65788]: WARNING openstack [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1442.593098] env[65788]: WARNING neutronclient.v2_0.client [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1442.593802] env[65788]: WARNING openstack [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1442.594153] env[65788]: WARNING openstack [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1442.675175] env[65788]: DEBUG nova.network.neutron [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updated VIF entry in instance network info cache for port bab8becc-f8fa-4a74-abbd-677bfd2508f1. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1442.675579] env[65788]: DEBUG nova.network.neutron [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [{"id": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "address": "fa:16:3e:27:d3:e2", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbab8becc-f8", "ovs_interfaceid": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1442.769789] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]528cb565-fe43-86fb-3e52-dca0bbcc6656, 'name': SearchDatastore_Task, 'duration_secs': 0.010632} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1442.770066] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1442.770308] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1442.770546] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1442.770690] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1442.770867] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1442.771155] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30fa5de5-2091-4204-b2e2-9993a7ae8be1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.780612] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1442.780787] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1442.781522] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16b4ea54-1b30-4c17-808f-02f8fb2f7987 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1442.787062] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1442.787062] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525d2520-b556-5fb8-d403-d153e7bb9ce2" [ 1442.787062] env[65788]: _type = "Task" [ 1442.787062] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1442.794994] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525d2520-b556-5fb8-d403-d153e7bb9ce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.179138] env[65788]: DEBUG oslo_concurrency.lockutils [req-f4933bdf-9d78-41ff-a922-8358df7d2144 req-4b8ec6ed-b4c6-494c-94dc-3bfd69da262b service nova] Releasing lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1443.298840] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]525d2520-b556-5fb8-d403-d153e7bb9ce2, 'name': SearchDatastore_Task, 'duration_secs': 0.009069} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.299682] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cdf8869-244d-45f7-afb8-778ba14516f7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.305603] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1443.305603] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d4d9c7-aa13-a0bf-1f1e-68ba936d027a" [ 1443.305603] env[65788]: _type = "Task" [ 1443.305603] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1443.313961] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d4d9c7-aa13-a0bf-1f1e-68ba936d027a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1443.324630] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1443.816642] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52d4d9c7-aa13-a0bf-1f1e-68ba936d027a, 'name': SearchDatastore_Task, 'duration_secs': 0.009802} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1443.816891] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1443.817200] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 1abcfa65-a4e9-452d-8be3-33724601be9f/1abcfa65-a4e9-452d-8be3-33724601be9f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1443.817471] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a7964bd-1d12-490d-b1f7-7ef962edf1c4 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1443.824727] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1443.824727] env[65788]: value = "task-4663476" [ 1443.824727] env[65788]: _type = "Task" [ 1443.824727] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1443.833434] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663476, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.335784] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663476, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471293} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.336247] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] 1abcfa65-a4e9-452d-8be3-33724601be9f/1abcfa65-a4e9-452d-8be3-33724601be9f.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1444.336547] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1444.336658] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ffaa007-c941-4dfa-aeaf-c25188930d9c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.344587] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1444.344587] env[65788]: value = "task-4663477" [ 1444.344587] env[65788]: _type = "Task" [ 1444.344587] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.353938] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1444.855047] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063006} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1444.855305] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1444.855949] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a6dcc6-b5e5-4d17-ae58-17379545d6ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.879615] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Reconfiguring VM instance instance-0000007e to attach disk [datastore2] 1abcfa65-a4e9-452d-8be3-33724601be9f/1abcfa65-a4e9-452d-8be3-33724601be9f.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1444.879917] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ad5888d-184a-4b87-a6a4-5063b9c76d5f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1444.900386] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1444.900386] env[65788]: value = "task-4663478" [ 1444.900386] env[65788]: _type = "Task" [ 1444.900386] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1444.908540] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663478, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.411136] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663478, 'name': ReconfigVM_Task, 'duration_secs': 0.362709} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.411511] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Reconfigured VM instance instance-0000007e to attach disk [datastore2] 1abcfa65-a4e9-452d-8be3-33724601be9f/1abcfa65-a4e9-452d-8be3-33724601be9f.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1445.412089] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dccb0c20-d4b6-4c40-86ca-8d71003c7c5e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.418685] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1445.418685] env[65788]: value = "task-4663479" [ 1445.418685] env[65788]: _type = "Task" [ 1445.418685] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.426559] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663479, 'name': Rename_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1445.638494] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1445.638678] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Cleaning up deleted instances with incomplete migration {{(pid=65788) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11954}} [ 1445.930650] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663479, 'name': Rename_Task, 'duration_secs': 0.141567} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1445.930928] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1445.931221] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-102a4281-9f36-4074-b9a2-c46894fc19ec {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1445.938833] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1445.938833] env[65788]: value = "task-4663480" [ 1445.938833] env[65788]: _type = "Task" [ 1445.938833] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1445.947405] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663480, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1446.449684] env[65788]: DEBUG oslo_vmware.api [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663480, 'name': PowerOnVM_Task, 'duration_secs': 0.426287} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1446.450131] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1446.450131] env[65788]: INFO nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Took 6.79 seconds to spawn the instance on the hypervisor. [ 1446.450319] env[65788]: DEBUG nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1446.451099] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968273fa-b8df-4f34-b92c-30980345c254 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1446.968607] env[65788]: INFO nova.compute.manager [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Took 11.58 seconds to build instance. [ 1447.428896] env[65788]: DEBUG nova.compute.manager [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Received event network-changed-bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1447.429048] env[65788]: DEBUG nova.compute.manager [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Refreshing instance network info cache due to event network-changed-bab8becc-f8fa-4a74-abbd-677bfd2508f1. {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11824}} [ 1447.429190] env[65788]: DEBUG oslo_concurrency.lockutils [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Acquiring lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1447.429339] env[65788]: DEBUG oslo_concurrency.lockutils [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Acquired lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1447.429494] env[65788]: DEBUG nova.network.neutron [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Refreshing network info cache for port bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2093}} [ 1447.470566] env[65788]: DEBUG oslo_concurrency.lockutils [None req-31bb9192-1d00-4d07-843a-9225a0f3f043 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.085s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1447.932592] env[65788]: WARNING neutronclient.v2_0.client [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1447.933315] env[65788]: WARNING openstack [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1447.933769] env[65788]: WARNING openstack [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1448.042940] env[65788]: WARNING openstack [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1448.043363] env[65788]: WARNING openstack [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1448.102930] env[65788]: WARNING neutronclient.v2_0.client [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1448.103639] env[65788]: WARNING openstack [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1448.103988] env[65788]: WARNING openstack [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1448.182145] env[65788]: DEBUG nova.network.neutron [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updated VIF entry in instance network info cache for port bab8becc-f8fa-4a74-abbd-677bfd2508f1. {{(pid=65788) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3599}} [ 1448.182560] env[65788]: DEBUG nova.network.neutron [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [{"id": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "address": "fa:16:3e:27:d3:e2", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbab8becc-f8", "ovs_interfaceid": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1448.685384] env[65788]: DEBUG oslo_concurrency.lockutils [req-6e29e1ff-0a08-485c-b229-ea1c543cfcb4 req-e3b7547f-0490-4060-804a-4b505b1c5455 service nova] Releasing lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1449.140871] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1449.141162] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Cleaning up deleted instances {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11916}} [ 1449.652848] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] There are 28 instances to clean {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11925}} [ 1449.653091] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 2f38c99d-0e4c-4c92-98fe-3af4fd9a79a0] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1450.156553] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: cc9be8c2-a6a8-4b33-9b15-e9ee58e43304] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1450.659565] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 16dfc544-45f3-486b-bc3d-86836382b07b] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1451.163438] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 4abbb248-12f6-4612-b6ea-01d1ef2104d7] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1451.667352] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 67934662-1b06-47b4-9fdf-d8f6ba17b86c] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1452.170609] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 222d1456-baad-42ee-bdef-020fb66ebe41] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1452.674492] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: ba0f0e87-fd5b-463a-bf6c-13394f2cb5e0] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1453.178029] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 152b3854-f2c6-471b-ac89-b32c51cfb762] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1453.681245] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: d15953f8-f74a-422a-867d-fa835bb479ff] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1454.184716] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 0636e867-8214-473e-9171-af66a53818a7] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1454.688017] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 31853be1-f03d-4cd9-91d7-e9542dc1cadc] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1455.191657] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 22cd2208-ebb5-401a-a06d-3dd6cd591869] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1455.695556] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f8c57215-ade6-424b-be8c-075998a764af] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1456.199602] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: f59d609a-e5bb-4202-ac05-d486bbf9e05b] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1456.703447] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: bdd09116-31de-491d-b129-c117f898881e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1457.207709] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 84a4e3bf-76cf-4699-8996-230c227cc1f8] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1457.710868] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 50c70e36-a3dc-4e8c-8fe6-74039d1828bf] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1458.214635] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 5621bbe8-131f-4266-9109-a3be445310a0] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1458.718251] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 6012182d-8b4e-4379-949f-d3a26211699b] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1459.222787] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 27ed2344-c42b-46bc-b51a-20821f67cbf0] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1459.726166] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 9946cef4-b8c7-41b4-9299-eace1929bf3d] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1460.230088] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: b9d8a32e-1f0f-405c-9543-9e53046d248e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1460.733644] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 475291c8-a16d-4980-9ef4-7ed6a4dc382e] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1461.236883] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 1f4769a0-f6d4-49eb-b012-bf5bdf6b11a7] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1461.739978] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 18be7322-3359-49c9-a181-f9228d5c16d7] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1462.244150] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 9d2f0d5e-793d-4790-ad31-be3d372be4fa] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1462.748624] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: 1a701758-a056-4948-9069-2a7168f2cc8c] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1463.252270] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] [instance: efe1048b-50e9-4add-910a-607a95759c7a] Instance has had 0 of 5 cleanup attempts {{(pid=65788) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11929}} [ 1484.829545] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "1abcfa65-a4e9-452d-8be3-33724601be9f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1484.829942] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1484.829987] env[65788]: DEBUG nova.compute.manager [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1484.830956] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630b6bb0-4629-4213-91df-31c090b3d4ce {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.838689] env[65788]: DEBUG nova.compute.manager [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=65788) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3442}} [ 1484.839345] env[65788]: DEBUG nova.objects.instance [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'flavor' on Instance uuid 1abcfa65-a4e9-452d-8be3-33724601be9f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1485.849325] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1485.849775] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdbb1236-a07f-4902-b8ed-61ea1d4117a7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1485.859021] env[65788]: DEBUG oslo_vmware.api [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1485.859021] env[65788]: value = "task-4663481" [ 1485.859021] env[65788]: _type = "Task" [ 1485.859021] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1485.868174] env[65788]: DEBUG oslo_vmware.api [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663481, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1486.369294] env[65788]: DEBUG oslo_vmware.api [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663481, 'name': PowerOffVM_Task, 'duration_secs': 0.184826} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1486.369598] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1486.369747] env[65788]: DEBUG nova.compute.manager [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1486.370527] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fc0604-b00b-485c-91a6-00aae4d21d62 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.881863] env[65788]: DEBUG oslo_concurrency.lockutils [None req-3f210375-f116-437e-b0ac-a2068f8d12be tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.052s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1487.355991] env[65788]: DEBUG nova.objects.instance [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'flavor' on Instance uuid 1abcfa65-a4e9-452d-8be3-33724601be9f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1487.861500] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1487.861734] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1487.861840] env[65788]: DEBUG nova.network.neutron [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1487.862048] env[65788]: DEBUG nova.objects.instance [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'info_cache' on Instance uuid 1abcfa65-a4e9-452d-8be3-33724601be9f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1488.365993] env[65788]: DEBUG nova.objects.base [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Object Instance<1abcfa65-a4e9-452d-8be3-33724601be9f> lazy-loaded attributes: flavor,info_cache {{(pid=65788) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1488.868807] env[65788]: WARNING neutronclient.v2_0.client [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1488.869789] env[65788]: WARNING openstack [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1488.870318] env[65788]: WARNING openstack [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1489.012172] env[65788]: WARNING openstack [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1489.012610] env[65788]: WARNING openstack [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1489.089187] env[65788]: WARNING neutronclient.v2_0.client [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1489.089882] env[65788]: WARNING openstack [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1489.090241] env[65788]: WARNING openstack [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1489.166960] env[65788]: DEBUG nova.network.neutron [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [{"id": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "address": "fa:16:3e:27:d3:e2", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbab8becc-f8", "ovs_interfaceid": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1489.670275] env[65788]: DEBUG oslo_concurrency.lockutils [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1490.676378] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1490.676822] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55315ca2-d401-4534-b95f-b40b60bcdc1c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.684873] env[65788]: DEBUG oslo_vmware.api [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1490.684873] env[65788]: value = "task-4663482" [ 1490.684873] env[65788]: _type = "Task" [ 1490.684873] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.693905] env[65788]: DEBUG oslo_vmware.api [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.195253] env[65788]: DEBUG oslo_vmware.api [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663482, 'name': PowerOnVM_Task, 'duration_secs': 0.361494} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.195522] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1491.195707] env[65788]: DEBUG nova.compute.manager [None req-b308e54d-53bf-4f4c-8cb6-6ffff4337c99 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1491.196493] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b4ee35-c5c8-45b1-9b15-414bd20b4d5c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.125540] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccf022f-b802-43bd-8624-4097ca11b4b0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.132721] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Suspending the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1492.132993] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1b37a2aa-cb4b-4050-847c-bba332b8da75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.142310] env[65788]: DEBUG oslo_vmware.api [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1492.142310] env[65788]: value = "task-4663483" [ 1492.142310] env[65788]: _type = "Task" [ 1492.142310] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1492.152317] env[65788]: DEBUG oslo_vmware.api [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663483, 'name': SuspendVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1492.653428] env[65788]: DEBUG oslo_vmware.api [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663483, 'name': SuspendVM_Task} progress is 66%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.153276] env[65788]: DEBUG oslo_vmware.api [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663483, 'name': SuspendVM_Task, 'duration_secs': 0.606622} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.153683] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Suspended the VM {{(pid=65788) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1493.153732] env[65788]: DEBUG nova.compute.manager [None req-8be34196-1485-466e-b0b9-4bd6e0221e62 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1493.154511] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9326a947-d617-47dd-965b-716f35fbd97c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.487296] env[65788]: INFO nova.compute.manager [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Resuming [ 1494.487954] env[65788]: DEBUG nova.objects.instance [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'flavor' on Instance uuid 1abcfa65-a4e9-452d-8be3-33724601be9f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1495.998678] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1495.999116] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquired lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1495.999116] env[65788]: DEBUG nova.network.neutron [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1496.502380] env[65788]: WARNING neutronclient.v2_0.client [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1496.503568] env[65788]: WARNING openstack [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1496.503882] env[65788]: WARNING openstack [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1496.658370] env[65788]: WARNING openstack [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1496.658790] env[65788]: WARNING openstack [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1496.722764] env[65788]: WARNING neutronclient.v2_0.client [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1496.723527] env[65788]: WARNING openstack [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1496.723902] env[65788]: WARNING openstack [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1496.802946] env[65788]: DEBUG nova.network.neutron [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [{"id": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "address": "fa:16:3e:27:d3:e2", "network": {"id": "0f261337-bbc6-42ba-afe6-3b3287aeb620", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-2006189981-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69c8fb9f477e45068cd7c1e4defb64a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbab8becc-f8", "ovs_interfaceid": "bab8becc-f8fa-4a74-abbd-677bfd2508f1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1497.305592] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Releasing lock "refresh_cache-1abcfa65-a4e9-452d-8be3-33724601be9f" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1497.306694] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd704d2f-6428-4d5b-8a60-55bd60d40309 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.313969] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Resuming the VM {{(pid=65788) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1497.314242] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e9ecf1f-eaed-41ac-b733-6a3df100f861 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1497.323241] env[65788]: DEBUG oslo_vmware.api [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1497.323241] env[65788]: value = "task-4663484" [ 1497.323241] env[65788]: _type = "Task" [ 1497.323241] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1497.332022] env[65788]: DEBUG oslo_vmware.api [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663484, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1497.841067] env[65788]: DEBUG oslo_vmware.api [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663484, 'name': PowerOnVM_Task} progress is 100%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1498.335548] env[65788]: DEBUG oslo_vmware.api [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663484, 'name': PowerOnVM_Task, 'duration_secs': 0.533019} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1498.335991] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Resumed the VM {{(pid=65788) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1498.336070] env[65788]: DEBUG nova.compute.manager [None req-5b55684c-8f86-4ad4-a869-d4f03bf1e8f5 tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1498.337016] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2e5a9e-ccd0-45c7-a95d-f5d9bb446baf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.231177] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "1abcfa65-a4e9-452d-8be3-33724601be9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1499.231446] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1499.231661] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "1abcfa65-a4e9-452d-8be3-33724601be9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1499.231884] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1499.232081] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1499.234861] env[65788]: INFO nova.compute.manager [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Terminating instance [ 1499.738462] env[65788]: DEBUG nova.compute.manager [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1499.738880] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1499.739628] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d9fc9d-ddf3-42bd-a4fc-f80bf049e754 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.748473] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1499.748749] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-781b0aa9-e682-4471-8fc9-5a400a991377 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1499.756431] env[65788]: DEBUG oslo_vmware.api [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1499.756431] env[65788]: value = "task-4663485" [ 1499.756431] env[65788]: _type = "Task" [ 1499.756431] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1499.767831] env[65788]: DEBUG oslo_vmware.api [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663485, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.267282] env[65788]: DEBUG oslo_vmware.api [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663485, 'name': PowerOffVM_Task, 'duration_secs': 0.21268} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.267573] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1500.267742] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1500.268010] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0986422f-abc1-47f6-8d38-a620966ec704 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.345558] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1500.345812] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1500.345989] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleting the datastore file [datastore2] 1abcfa65-a4e9-452d-8be3-33724601be9f {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1500.346284] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b10c8223-ae29-4073-9d0f-fc1060df16ac {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1500.355228] env[65788]: DEBUG oslo_vmware.api [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for the task: (returnval){ [ 1500.355228] env[65788]: value = "task-4663487" [ 1500.355228] env[65788]: _type = "Task" [ 1500.355228] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1500.364134] env[65788]: DEBUG oslo_vmware.api [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663487, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1500.865532] env[65788]: DEBUG oslo_vmware.api [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Task: {'id': task-4663487, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135426} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1500.865918] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1500.865918] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1500.866067] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1500.866248] env[65788]: INFO nova.compute.manager [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1500.866506] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1500.866730] env[65788]: DEBUG nova.compute.manager [-] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1500.866822] env[65788]: DEBUG nova.network.neutron [-] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1500.867073] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1500.867592] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1500.867859] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1500.904829] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1501.419117] env[65788]: DEBUG nova.compute.manager [req-acedd0ee-c1cb-4efa-8fe1-01e9e7e2bf47 req-99c8e198-e18c-41a8-ab0b-e7b823639938 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Received event network-vif-deleted-bab8becc-f8fa-4a74-abbd-677bfd2508f1 {{(pid=65788) external_instance_event /opt/stack/nova/nova/compute/manager.py:11819}} [ 1501.419355] env[65788]: INFO nova.compute.manager [req-acedd0ee-c1cb-4efa-8fe1-01e9e7e2bf47 req-99c8e198-e18c-41a8-ab0b-e7b823639938 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Neutron deleted interface bab8becc-f8fa-4a74-abbd-677bfd2508f1; detaching it from the instance and deleting it from the info cache [ 1501.419527] env[65788]: DEBUG nova.network.neutron [req-acedd0ee-c1cb-4efa-8fe1-01e9e7e2bf47 req-99c8e198-e18c-41a8-ab0b-e7b823639938 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1501.850087] env[65788]: DEBUG nova.network.neutron [-] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1501.924041] env[65788]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47aaec39-08fb-4e00-8e86-941ab22a5c68 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.933948] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dfa86c-c6b8-4b6d-b6a1-0d8dc11c8d27 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1501.959196] env[65788]: DEBUG nova.compute.manager [req-acedd0ee-c1cb-4efa-8fe1-01e9e7e2bf47 req-99c8e198-e18c-41a8-ab0b-e7b823639938 service nova] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Detach interface failed, port_id=bab8becc-f8fa-4a74-abbd-677bfd2508f1, reason: Instance 1abcfa65-a4e9-452d-8be3-33724601be9f could not be found. {{(pid=65788) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1502.352671] env[65788]: INFO nova.compute.manager [-] [instance: 1abcfa65-a4e9-452d-8be3-33724601be9f] Took 1.49 seconds to deallocate network for instance. [ 1502.859264] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1502.859573] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1502.859799] env[65788]: DEBUG nova.objects.instance [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lazy-loading 'resources' on Instance uuid 1abcfa65-a4e9-452d-8be3-33724601be9f {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1503.380706] env[65788]: DEBUG nova.scheduler.client.report [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Refreshing inventories for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1503.396293] env[65788]: DEBUG nova.scheduler.client.report [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Updating ProviderTree inventory for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1503.396520] env[65788]: DEBUG nova.compute.provider_tree [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Updating inventory in ProviderTree for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1503.407727] env[65788]: DEBUG nova.scheduler.client.report [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Refreshing aggregate associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, aggregates: None {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1503.426313] env[65788]: DEBUG nova.scheduler.client.report [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Refreshing trait associations for resource provider 3a6417f0-b161-4086-8a0e-1cb139eae377, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=65788) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1503.454185] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28dce42-e203-440f-bfc3-8931a57c3e11 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.462452] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4590e2d3-7a23-4775-881a-6ab974b8de4b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.492954] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a8b45d-5ecd-450c-9e80-1a1d5c2a013a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.501166] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c646bd-23c2-47d5-a043-bfae68b6f79f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1503.516446] env[65788]: DEBUG nova.compute.provider_tree [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1504.020505] env[65788]: DEBUG nova.scheduler.client.report [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1504.525816] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1504.548517] env[65788]: INFO nova.scheduler.client.report [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Deleted allocations for instance 1abcfa65-a4e9-452d-8be3-33724601be9f [ 1505.057649] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c62ea071-af96-40e2-a09c-ba9e1a1f6f3d tempest-ServerActionsTestJSON-316964864 tempest-ServerActionsTestJSON-316964864-project-member] Lock "1abcfa65-a4e9-452d-8be3-33724601be9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.826s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1508.252928] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.253357] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.368518] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "f2e16e72-53e9-4fc0-93e3-e163025f6888" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1508.368771] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "f2e16e72-53e9-4fc0-93e3-e163025f6888" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1508.759778] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.759991] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.760186] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.760351] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.760495] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.760641] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.761029] env[65788]: DEBUG nova.compute.manager [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=65788) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1508.761029] env[65788]: DEBUG oslo_service.periodic_task [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Running periodic task ComputeManager.update_available_resource {{(pid=65788) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1508.870985] env[65788]: DEBUG nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Starting instance... {{(pid=65788) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2476}} [ 1509.263665] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1509.264096] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1509.264203] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1509.264426] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=65788) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 1509.265791] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24ad5cf-5900-4c58-9691-58eff4dec89c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.274365] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c24f579-6c82-42e7-bec9-3fd38ff9e95c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.290908] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743076a4-c9a4-44f6-b524-eaff57dbf4fc {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.299378] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5b80d5-da5c-4cbf-86f1-cca6813e1bc7 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1509.330027] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180618MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=65788) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 1509.330239] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1509.330463] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1509.394936] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1510.855746] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Instance f2e16e72-53e9-4fc0-93e3-e163025f6888 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=65788) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 1510.856192] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 1510.856192] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=65788) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 1510.882981] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b0aa1e-3055-4ff0-afeb-c5871d57aeb9 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1510.890766] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4eb4da-9213-4fc0-b0f8-1ee91d525719 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1510.922233] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347ad814-a7e4-4328-9e6c-6966063ad6d6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1510.930363] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b9efcd-b4ff-4fe3-8cf3-a2e201c3238c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1510.943931] env[65788]: DEBUG nova.compute.provider_tree [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1511.447035] env[65788]: DEBUG nova.scheduler.client.report [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1511.952326] env[65788]: DEBUG nova.compute.resource_tracker [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=65788) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 1511.952667] env[65788]: DEBUG oslo_concurrency.lockutils [None req-51f2de1d-a45d-4c28-817a-39777e7ea45e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.622s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1511.952859] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.558s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1511.954517] env[65788]: INFO nova.compute.claims [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1512.995121] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611b2435-5812-4558-aa64-32747e203dc3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.003662] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c295f14-7e5c-490d-a943-8ffd4ea8f9a3 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.036950] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99290385-b56f-4c87-8bac-718fdd39c2de {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.045734] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fada6f1-337b-475c-85a6-4e602cfcfeef {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1513.060889] env[65788]: DEBUG nova.compute.provider_tree [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1513.563900] env[65788]: DEBUG nova.scheduler.client.report [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1514.069507] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.116s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1514.070261] env[65788]: DEBUG nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Start building networks asynchronously for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2874}} [ 1514.575822] env[65788]: DEBUG nova.compute.utils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Using /dev/sd instead of None {{(pid=65788) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1514.577769] env[65788]: DEBUG nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Not allocating networking since 'none' was specified. {{(pid=65788) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2016}} [ 1515.080064] env[65788]: DEBUG nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Start building block device mappings for instance. {{(pid=65788) _build_resources /opt/stack/nova/nova/compute/manager.py:2909}} [ 1516.090642] env[65788]: DEBUG nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Start spawning the instance on the hypervisor. {{(pid=65788) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2682}} [ 1516.116868] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1516.117159] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1516.117321] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1516.117506] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1516.117650] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1516.117842] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1516.118071] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1516.118236] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1516.118407] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1516.118571] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1516.118751] env[65788]: DEBUG nova.virt.hardware [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1516.119650] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc462aab-62ba-45db-b30f-cea1849388cd {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.127947] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e4665b-9aa6-4778-8d92-9331060d4695 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.141740] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1516.147198] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Creating folder: Project (2e672aef8e0440ecb0d62b8176481f03). Parent ref: group-v910111. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1516.147480] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff5621cf-4167-43e4-a72e-5703c6545363 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.159602] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Created folder: Project (2e672aef8e0440ecb0d62b8176481f03) in parent group-v910111. [ 1516.159796] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Creating folder: Instances. Parent ref: group-v910456. {{(pid=65788) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1516.160030] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3677ec9-1109-4ee0-94b3-94cd842f1f6e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.171013] env[65788]: INFO nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Created folder: Instances in parent group-v910456. [ 1516.171244] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1516.171457] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1516.171673] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d316f43e-9d3a-4b06-b224-e86fec3a635d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1516.188370] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1516.188370] env[65788]: value = "task-4663490" [ 1516.188370] env[65788]: _type = "Task" [ 1516.188370] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1516.196509] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663490, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1516.701050] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663490, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1517.200195] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663490, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1517.700248] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663490, 'name': CreateVM_Task, 'duration_secs': 1.260002} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1517.700446] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1517.700933] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1517.701158] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1517.701596] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1517.701885] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92cee8cd-3487-4fc5-8ad5-88560bebc27c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.707525] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1517.707525] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52423b12-d133-2b56-5772-f9d03cfe567a" [ 1517.707525] env[65788]: _type = "Task" [ 1517.707525] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1517.715464] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52423b12-d133-2b56-5772-f9d03cfe567a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1518.217936] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52423b12-d133-2b56-5772-f9d03cfe567a, 'name': SearchDatastore_Task, 'duration_secs': 0.011144} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1518.218391] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1518.218539] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1518.218773] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1518.218920] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1518.219127] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1518.219386] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0ef65aa-5856-46cc-ae04-2240428b7d81 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.229320] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1518.229540] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1518.230349] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8079a21-c0b2-4bae-98c3-3b7966f46857 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.236512] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1518.236512] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524fdefa-21b6-1290-7b82-2f971e4fb0ae" [ 1518.236512] env[65788]: _type = "Task" [ 1518.236512] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1518.244951] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524fdefa-21b6-1290-7b82-2f971e4fb0ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1518.747584] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524fdefa-21b6-1290-7b82-2f971e4fb0ae, 'name': SearchDatastore_Task, 'duration_secs': 0.008863} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1518.748358] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7846c25e-0586-4617-bd9f-8ba4484b3e61 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1518.754172] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1518.754172] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52923841-5e0e-0e70-dc7a-da138ebcf95d" [ 1518.754172] env[65788]: _type = "Task" [ 1518.754172] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1518.762102] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52923841-5e0e-0e70-dc7a-da138ebcf95d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1519.264979] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]52923841-5e0e-0e70-dc7a-da138ebcf95d, 'name': SearchDatastore_Task, 'duration_secs': 0.00951} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1519.265675] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1519.265675] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1519.265869] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f68d220e-f1ba-42c9-b970-302f8177cf7a {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.277280] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1519.277280] env[65788]: value = "task-4663491" [ 1519.277280] env[65788]: _type = "Task" [ 1519.277280] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1519.285865] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1519.787618] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663491, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442573} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1519.788060] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1519.789975] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1519.789975] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4546880-1fb0-4362-9591-a13b755e639b {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.795558] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1519.795558] env[65788]: value = "task-4663492" [ 1519.795558] env[65788]: _type = "Task" [ 1519.795558] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1519.804885] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663492, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1520.305849] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663492, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058575} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1520.306226] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1520.306959] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cf786e-7912-4875-8429-42780a6a5fe1 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.327180] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Reconfiguring VM instance instance-0000007f to attach disk [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1520.327446] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-423d0022-3fd7-4cca-b0bb-9b2fb597598f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.348539] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1520.348539] env[65788]: value = "task-4663493" [ 1520.348539] env[65788]: _type = "Task" [ 1520.348539] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1520.359687] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663493, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1520.859182] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663493, 'name': ReconfigVM_Task, 'duration_secs': 0.302906} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1520.859527] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Reconfigured VM instance instance-0000007f to attach disk [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1520.860282] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d01d505a-8080-4d5e-a932-68bcc9e16f77 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.867617] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1520.867617] env[65788]: value = "task-4663494" [ 1520.867617] env[65788]: _type = "Task" [ 1520.867617] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1520.876369] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663494, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1521.377506] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663494, 'name': Rename_Task, 'duration_secs': 0.124535} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1521.378057] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1521.378057] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8fabd83-dca2-48f3-b306-9caf54e57e75 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.384737] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1521.384737] env[65788]: value = "task-4663495" [ 1521.384737] env[65788]: _type = "Task" [ 1521.384737] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1521.392724] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663495, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1521.895824] env[65788]: DEBUG oslo_vmware.api [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663495, 'name': PowerOnVM_Task, 'duration_secs': 0.399662} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1521.896107] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1521.896312] env[65788]: INFO nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Took 5.81 seconds to spawn the instance on the hypervisor. [ 1521.896498] env[65788]: DEBUG nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1521.897310] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b4bfb6-dd0a-4c3b-95cf-ae11e0af1485 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.415693] env[65788]: INFO nova.compute.manager [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Took 13.04 seconds to build instance. [ 1522.917863] env[65788]: DEBUG oslo_concurrency.lockutils [None req-5e3a7261-d4f9-419a-856b-e7f48311b011 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "f2e16e72-53e9-4fc0-93e3-e163025f6888" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.549s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1523.698037] env[65788]: INFO nova.compute.manager [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Rebuilding instance [ 1523.736955] env[65788]: DEBUG nova.compute.manager [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1523.737851] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302c1a93-4659-411c-a912-66f3ab9ea5f6 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.751905] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1524.752345] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edff6d78-c54a-4081-8437-6a555b31a004 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.762081] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1524.762081] env[65788]: value = "task-4663496" [ 1524.762081] env[65788]: _type = "Task" [ 1524.762081] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1524.770815] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.272051] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663496, 'name': PowerOffVM_Task, 'duration_secs': 0.112973} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1525.272332] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1525.272981] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1525.273839] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa984eb-6d6b-4510-8dd3-1f1c2aa2a9f5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.280752] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1525.280979] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99377514-89e9-48d5-accb-3130a0169d70 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.308180] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1525.308388] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1525.308568] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Deleting the datastore file [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1525.308918] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7497bb8d-46d3-40b8-a0c5-4ffdca030af5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1525.316190] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1525.316190] env[65788]: value = "task-4663498" [ 1525.316190] env[65788]: _type = "Task" [ 1525.316190] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1525.324884] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663498, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1525.826335] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663498, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097526} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1525.826814] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1525.826814] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1525.826933] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1526.863523] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-11-21T13:43:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-11-21T13:43:27Z,direct_url=,disk_format='vmdk',id=d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='a305d1aa178f40f2a9e68e92449a002b',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-11-21T13:43:27Z,virtual_size=,visibility=), allow threads: False {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 1526.863829] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Flavor limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1526.863934] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Image limits 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 1526.864132] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Flavor pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1526.864278] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Image pref 0:0:0 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 1526.864422] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=65788) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 1526.864627] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1526.864782] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 1526.864941] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Got 1 possible topologies {{(pid=65788) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 1526.865116] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 1526.865294] env[65788]: DEBUG nova.virt.hardware [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=65788) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 1526.866205] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdfb175-e3f3-4b9c-a620-b5973f9ca7eb {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1526.875775] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2aa3137-1226-4b42-a9ad-40bb9f86568e {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1526.889866] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Instance VIF info [] {{(pid=65788) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1526.895465] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1526.895732] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Creating VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1526.895946] env[65788]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2976c89-3bc2-469d-9871-08a72a953d23 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1526.912046] env[65788]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1526.912046] env[65788]: value = "task-4663499" [ 1526.912046] env[65788]: _type = "Task" [ 1526.912046] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1526.919872] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663499, 'name': CreateVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1527.422611] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663499, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1527.923288] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663499, 'name': CreateVM_Task} progress is 99%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1528.424901] env[65788]: DEBUG oslo_vmware.api [-] Task: {'id': task-4663499, 'name': CreateVM_Task, 'duration_secs': 1.257962} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1528.425123] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Created VM on the ESX host {{(pid=65788) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1528.425552] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1528.425709] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1528.426047] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1528.426313] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e06ed3b0-cd89-4a10-9c76-a6a9c1fa36cf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.432059] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1528.432059] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5257b3a8-fb6d-f9e0-09ca-ca982e2b1edc" [ 1528.432059] env[65788]: _type = "Task" [ 1528.432059] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1528.440937] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5257b3a8-fb6d-f9e0-09ca-ca982e2b1edc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1528.945651] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]5257b3a8-fb6d-f9e0-09ca-ca982e2b1edc, 'name': SearchDatastore_Task, 'duration_secs': 0.012256} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1528.946125] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1528.946370] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Processing image d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b {{(pid=65788) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1528.946689] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1528.946907] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1528.947171] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1528.947520] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de889847-b454-4c4a-ace8-870ce92a4475 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.957591] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=65788) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1528.957833] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=65788) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1528.958841] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f8ff0ca-c2e7-47e5-b2d4-9dd71b3462e5 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.965764] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1528.965764] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524da83e-3630-23d5-b615-8db1b8e9ca2a" [ 1528.965764] env[65788]: _type = "Task" [ 1528.965764] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1528.976565] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524da83e-3630-23d5-b615-8db1b8e9ca2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1529.478264] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]524da83e-3630-23d5-b615-8db1b8e9ca2a, 'name': SearchDatastore_Task, 'duration_secs': 0.009138} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1529.479103] env[65788]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9657020-3f6e-4201-bcf5-510811eff6be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.484513] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1529.484513] env[65788]: value = "session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b5e2a-e752-ad02-9782-42bb3730f01d" [ 1529.484513] env[65788]: _type = "Task" [ 1529.484513] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1529.492712] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b5e2a-e752-ad02-9782-42bb3730f01d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1529.995098] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': session[52a6be1d-aa7d-be7d-8c0d-3125c16d35e3]522b5e2a-e752-ad02-9782-42bb3730f01d, 'name': SearchDatastore_Task, 'duration_secs': 0.00918} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1529.995489] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1529.995614] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1529.995883] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33ccd39f-4d39-423d-84be-923ecace862d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.003786] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1530.003786] env[65788]: value = "task-4663500" [ 1530.003786] env[65788]: _type = "Task" [ 1530.003786] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1530.011990] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1530.514034] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443824} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1530.514323] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b/d5a7b45a-a1f3-4d64-bb88-d3ab995bc18b.vmdk to [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk {{(pid=65788) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1530.514533] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Extending root virtual disk to 1048576 {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1530.514812] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7069241-874b-4eee-80a2-ff1be1a6a6b0 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.521285] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1530.521285] env[65788]: value = "task-4663501" [ 1530.521285] env[65788]: _type = "Task" [ 1530.521285] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1530.529558] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1531.031728] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058115} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1531.032302] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Extended root virtual disk {{(pid=65788) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1531.032806] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abd5c0f-231d-4cb6-879a-4172a8dac37c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.052885] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Reconfiguring VM instance instance-0000007f to attach disk [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1531.053173] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35dc5ca5-24e6-46bc-b98a-4a4a20179276 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.072874] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1531.072874] env[65788]: value = "task-4663502" [ 1531.072874] env[65788]: _type = "Task" [ 1531.072874] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1531.080805] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663502, 'name': ReconfigVM_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1531.583949] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663502, 'name': ReconfigVM_Task, 'duration_secs': 0.303144} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1531.584261] env[65788]: DEBUG nova.virt.vmwareapi.volumeops [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Reconfigured VM instance instance-0000007f to attach disk [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888/f2e16e72-53e9-4fc0-93e3-e163025f6888.vmdk or device None with type sparse {{(pid=65788) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1531.584979] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d880f1b8-49a2-4f7c-9de6-b7c63f93230f {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1531.592083] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1531.592083] env[65788]: value = "task-4663503" [ 1531.592083] env[65788]: _type = "Task" [ 1531.592083] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1531.600930] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663503, 'name': Rename_Task} progress is 5%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1532.102957] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663503, 'name': Rename_Task, 'duration_secs': 0.128429} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1532.103406] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powering on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1532.103516] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff6640fc-8683-458c-abb6-3e90a543bced {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1532.111862] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1532.111862] env[65788]: value = "task-4663504" [ 1532.111862] env[65788]: _type = "Task" [ 1532.111862] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1532.119905] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1532.621654] env[65788]: DEBUG oslo_vmware.api [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663504, 'name': PowerOnVM_Task, 'duration_secs': 0.392634} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1532.621932] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powered on the VM {{(pid=65788) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1532.622148] env[65788]: DEBUG nova.compute.manager [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Checking state {{(pid=65788) _get_power_state /opt/stack/nova/nova/compute/manager.py:1829}} [ 1532.622935] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55aab3f-35c1-4173-aa73-3ba52b651c50 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1533.138972] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1533.139360] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1533.139360] env[65788]: DEBUG nova.objects.instance [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Trying to apply a migration context that does not seem to be set for this instance {{(pid=65788) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1533.471267] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "f2e16e72-53e9-4fc0-93e3-e163025f6888" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1533.471516] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "f2e16e72-53e9-4fc0-93e3-e163025f6888" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1533.471721] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "f2e16e72-53e9-4fc0-93e3-e163025f6888-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1533.471916] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "f2e16e72-53e9-4fc0-93e3-e163025f6888-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1533.472105] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "f2e16e72-53e9-4fc0-93e3-e163025f6888-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1533.474397] env[65788]: INFO nova.compute.manager [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Terminating instance [ 1533.979063] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "refresh_cache-f2e16e72-53e9-4fc0-93e3-e163025f6888" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1533.979063] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquired lock "refresh_cache-f2e16e72-53e9-4fc0-93e3-e163025f6888" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1533.979353] env[65788]: DEBUG nova.network.neutron [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Building network info cache for instance {{(pid=65788) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2096}} [ 1534.149879] env[65788]: DEBUG oslo_concurrency.lockutils [None req-c400f548-83fb-410a-990f-2ca395e324de tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1534.482487] env[65788]: WARNING neutronclient.v2_0.client [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1534.483202] env[65788]: WARNING openstack [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1534.483576] env[65788]: WARNING openstack [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1534.506746] env[65788]: DEBUG nova.network.neutron [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1534.569343] env[65788]: DEBUG nova.network.neutron [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1535.071923] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Releasing lock "refresh_cache-f2e16e72-53e9-4fc0-93e3-e163025f6888" {{(pid=65788) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1535.072393] env[65788]: DEBUG nova.compute.manager [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Start destroying the instance on the hypervisor. {{(pid=65788) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3202}} [ 1535.072595] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Destroying instance {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1535.073519] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8782a82-eafb-4271-932b-06024bc9e919 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1535.082436] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powering off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1535.082744] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c04a862c-99d9-431c-adb5-7ff3e7ab469d {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1535.089432] env[65788]: DEBUG oslo_vmware.api [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1535.089432] env[65788]: value = "task-4663505" [ 1535.089432] env[65788]: _type = "Task" [ 1535.089432] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1535.098436] env[65788]: DEBUG oslo_vmware.api [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663505, 'name': PowerOffVM_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1535.599758] env[65788]: DEBUG oslo_vmware.api [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663505, 'name': PowerOffVM_Task, 'duration_secs': 0.109672} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1535.600192] env[65788]: DEBUG nova.virt.vmwareapi.vm_util [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Powered off the VM {{(pid=65788) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1535.600234] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Unregistering the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1535.600466] env[65788]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6146516-7997-4ee8-9efb-f5fe4f00784c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1535.626966] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Unregistered the VM {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1535.627288] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Deleting contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1535.627479] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Deleting the datastore file [datastore2] f2e16e72-53e9-4fc0-93e3-e163025f6888 {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1535.627758] env[65788]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2a34ba9-0664-4458-ae89-f156c424f0bf {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1535.635081] env[65788]: DEBUG oslo_vmware.api [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for the task: (returnval){ [ 1535.635081] env[65788]: value = "task-4663507" [ 1535.635081] env[65788]: _type = "Task" [ 1535.635081] env[65788]: } to complete. {{(pid=65788) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1535.643323] env[65788]: DEBUG oslo_vmware.api [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1536.145299] env[65788]: DEBUG oslo_vmware.api [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Task: {'id': task-4663507, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102152} completed successfully. {{(pid=65788) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1536.145590] env[65788]: DEBUG nova.virt.vmwareapi.ds_util [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Deleted the datastore file {{(pid=65788) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1536.145735] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Deleted contents of the VM from datastore datastore2 {{(pid=65788) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1536.145910] env[65788]: DEBUG nova.virt.vmwareapi.vmops [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Instance destroyed {{(pid=65788) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1536.146101] env[65788]: INFO nova.compute.manager [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1536.146362] env[65788]: DEBUG oslo.service.backend._common.loopingcall [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=65788) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 1536.146555] env[65788]: DEBUG nova.compute.manager [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Deallocating network for instance {{(pid=65788) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2328}} [ 1536.146682] env[65788]: DEBUG nova.network.neutron [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] deallocate_for_instance() {{(pid=65788) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1886}} [ 1536.146925] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1536.147471] env[65788]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 1536.147730] env[65788]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 1536.169061] env[65788]: DEBUG nova.network.neutron [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Instance cache missing network info. {{(pid=65788) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3438}} [ 1536.169310] env[65788]: WARNING neutronclient.v2_0.client [-] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 1536.672432] env[65788]: DEBUG nova.network.neutron [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Updating instance_info_cache with network_info: [] {{(pid=65788) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 1537.175772] env[65788]: INFO nova.compute.manager [-] [instance: f2e16e72-53e9-4fc0-93e3-e163025f6888] Took 1.03 seconds to deallocate network for instance. [ 1537.682823] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1537.683243] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1537.683355] env[65788]: DEBUG nova.objects.instance [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lazy-loading 'resources' on Instance uuid f2e16e72-53e9-4fc0-93e3-e163025f6888 {{(pid=65788) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1538.216045] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df42cfa-2928-4f7d-be19-4ecb73f8742c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.223750] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e36993-a4d5-441d-a552-3a6af42823be {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.254088] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a09de59-c89f-4313-8a02-069ed29de20c {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.262332] env[65788]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07cac7c-b82c-4010-8de5-498da23ba072 {{(pid=65788) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.276487] env[65788]: DEBUG nova.compute.provider_tree [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Inventory has not changed in ProviderTree for provider: 3a6417f0-b161-4086-8a0e-1cb139eae377 {{(pid=65788) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1538.780225] env[65788]: DEBUG nova.scheduler.client.report [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Inventory has not changed for provider 3a6417f0-b161-4086-8a0e-1cb139eae377 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=65788) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1539.286226] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.603s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1539.324973] env[65788]: INFO nova.scheduler.client.report [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Deleted allocations for instance f2e16e72-53e9-4fc0-93e3-e163025f6888 [ 1539.836103] env[65788]: DEBUG oslo_concurrency.lockutils [None req-145d8ffa-5feb-4a9c-9edd-4c3d1395bf99 tempest-ServersListShow298Test-1628129887 tempest-ServersListShow298Test-1628129887-project-member] Lock "f2e16e72-53e9-4fc0-93e3-e163025f6888" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.364s {{(pid=65788) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}